text
stringlengths 493
2.1k
| inputs
dict | prediction
null | prediction_agent
null | annotation
stringclasses 2
values | annotation_agent
stringclasses 1
value | vectors
null | multi_label
bool 1
class | explanation
null | id
stringlengths 36
36
| metadata
null | status
stringclasses 2
values | event_timestamp
stringlengths 26
26
| metrics
dict | label
class label 2
classes |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
TITLE: Structural Similarity: When to Use Deep Generative Models on Imbalanced Image Dataset Augmentation
ABSTRACT: Improving the performance on an imbalanced training set is one of the main
challenges in nowadays Machine Learning. One way to augment and thus re-balance
the image dataset is through existing deep generative models, like
class-conditional Generative Adversarial Networks (cGAN) or Diffusion Models by
synthesizing images on each of the tail-class. Our experiments on imbalanced
image dataset classification show that, the validation accuracy improvement
with such re-balancing method is related to the image similarity between
different classes. Thus, to quantify this image dataset class similarity, we
propose a measurement called Super-Sub Class Structural Similarity
(SSIM-supSubCls) based on Structural Similarity (SSIM). A deep generative model
data augmentation classification (GM-augCls) pipeline is also provided to
verify this metric correlates with the accuracy enhancement. We further
quantify the relationship between them, discovering that the accuracy
improvement decays exponentially with respect to SSIM-supSubCls values. | {
"abstract": "Improving the performance on an imbalanced training set is one of the main\nchallenges in nowadays Machine Learning. One way to augment and thus re-balance\nthe image dataset is through existing deep generative models, like\nclass-conditional Generative Adversarial Networks (cGAN) or Diffusion Models by\nsynthesizing images on each of the tail-class. Our experiments on imbalanced\nimage dataset classification show that, the validation accuracy improvement\nwith such re-balancing method is related to the image similarity between\ndifferent classes. Thus, to quantify this image dataset class similarity, we\npropose a measurement called Super-Sub Class Structural Similarity\n(SSIM-supSubCls) based on Structural Similarity (SSIM). A deep generative model\ndata augmentation classification (GM-augCls) pipeline is also provided to\nverify this metric correlates with the accuracy enhancement. We further\nquantify the relationship between them, discovering that the accuracy\nimprovement decays exponentially with respect to SSIM-supSubCls values.",
"title": "Structural Similarity: When to Use Deep Generative Models on Imbalanced Image Dataset Augmentation",
"url": "http://arxiv.org/abs/2303.04854v1"
} | null | null | no_new_dataset | admin | null | false | null | 0a2c3242-5fcd-42cd-9482-6347fd9a06fa | null | Validated | 2023-10-04 15:19:51.880559 | {
"text_length": 1172
} | 1no_new_dataset
|
TITLE: Gender Bias in Text: Labeled Datasets and Lexicons
ABSTRACT: Language has a profound impact on our thoughts, perceptions, and conceptions
of gender roles. Gender-inclusive language is, therefore, a key tool to promote
social inclusion and contribute to achieving gender equality. Consequently,
detecting and mitigating gender bias in texts is instrumental in halting its
propagation and societal implications. However, there is a lack of gender bias
datasets and lexicons for automating the detection of gender bias using
supervised and unsupervised machine learning (ML) and natural language
processing (NLP) techniques. Therefore, the main contribution of this work is
to publicly provide labeled datasets and exhaustive lexicons by collecting,
annotating, and augmenting relevant sentences to facilitate the detection of
gender bias in English text. Towards this end, we present an updated version of
our previously proposed taxonomy by re-formalizing its structure, adding a new
bias type, and mapping each bias subtype to an appropriate detection
methodology. The released datasets and lexicons span multiple bias subtypes
including: Generic He, Generic She, Explicit Marking of Sex, and Gendered
Neologisms. We leveraged the use of word embedding models to further augment
the collected lexicons. | {
"abstract": "Language has a profound impact on our thoughts, perceptions, and conceptions\nof gender roles. Gender-inclusive language is, therefore, a key tool to promote\nsocial inclusion and contribute to achieving gender equality. Consequently,\ndetecting and mitigating gender bias in texts is instrumental in halting its\npropagation and societal implications. However, there is a lack of gender bias\ndatasets and lexicons for automating the detection of gender bias using\nsupervised and unsupervised machine learning (ML) and natural language\nprocessing (NLP) techniques. Therefore, the main contribution of this work is\nto publicly provide labeled datasets and exhaustive lexicons by collecting,\nannotating, and augmenting relevant sentences to facilitate the detection of\ngender bias in English text. Towards this end, we present an updated version of\nour previously proposed taxonomy by re-formalizing its structure, adding a new\nbias type, and mapping each bias subtype to an appropriate detection\nmethodology. The released datasets and lexicons span multiple bias subtypes\nincluding: Generic He, Generic She, Explicit Marking of Sex, and Gendered\nNeologisms. We leveraged the use of word embedding models to further augment\nthe collected lexicons.",
"title": "Gender Bias in Text: Labeled Datasets and Lexicons",
"url": "http://arxiv.org/abs/2201.08675v2"
} | null | null | new_dataset | admin | null | false | null | 859606d1-fd05-4ada-88ff-f2d4e1df1e01 | null | Validated | 2023-10-04 15:19:51.888725 | {
"text_length": 1326
} | 0new_dataset
|
TITLE: The ArtBench Dataset: Benchmarking Generative Models with Artworks
ABSTRACT: We introduce ArtBench-10, the first class-balanced, high-quality, cleanly
annotated, and standardized dataset for benchmarking artwork generation. It
comprises 60,000 images of artwork from 10 distinctive artistic styles, with
5,000 training images and 1,000 testing images per style. ArtBench-10 has
several advantages over previous artwork datasets. Firstly, it is
class-balanced while most previous artwork datasets suffer from the long tail
class distributions. Secondly, the images are of high quality with clean
annotations. Thirdly, ArtBench-10 is created with standardized data collection,
annotation, filtering, and preprocessing procedures. We provide three versions
of the dataset with different resolutions ($32\times32$, $256\times256$, and
original image size), formatted in a way that is easy to be incorporated by
popular machine learning frameworks. We also conduct extensive benchmarking
experiments using representative image synthesis models with ArtBench-10 and
present in-depth analysis. The dataset is available at
https://github.com/liaopeiyuan/artbench under a Fair Use license. | {
"abstract": "We introduce ArtBench-10, the first class-balanced, high-quality, cleanly\nannotated, and standardized dataset for benchmarking artwork generation. It\ncomprises 60,000 images of artwork from 10 distinctive artistic styles, with\n5,000 training images and 1,000 testing images per style. ArtBench-10 has\nseveral advantages over previous artwork datasets. Firstly, it is\nclass-balanced while most previous artwork datasets suffer from the long tail\nclass distributions. Secondly, the images are of high quality with clean\nannotations. Thirdly, ArtBench-10 is created with standardized data collection,\nannotation, filtering, and preprocessing procedures. We provide three versions\nof the dataset with different resolutions ($32\\times32$, $256\\times256$, and\noriginal image size), formatted in a way that is easy to be incorporated by\npopular machine learning frameworks. We also conduct extensive benchmarking\nexperiments using representative image synthesis models with ArtBench-10 and\npresent in-depth analysis. The dataset is available at\nhttps://github.com/liaopeiyuan/artbench under a Fair Use license.",
"title": "The ArtBench Dataset: Benchmarking Generative Models with Artworks",
"url": "http://arxiv.org/abs/2206.11404v1"
} | null | null | new_dataset | admin | null | false | null | 7e646a05-d5a1-426a-8b26-65595dc49b95 | null | Validated | 2023-10-04 15:19:51.885607 | {
"text_length": 1204
} | 0new_dataset
|
TITLE: SeFNet: Bridging Tabular Datasets with Semantic Feature Nets
ABSTRACT: Machine learning applications cover a wide range of predictive tasks in which
tabular datasets play a significant role. However, although they often address
similar problems, tabular datasets are typically treated as standalone tasks.
The possibilities of using previously solved problems are limited due to the
lack of structured contextual information about their features and the lack of
understanding of the relations between them. To overcome this limitation, we
propose a new approach called Semantic Feature Net (SeFNet), capturing the
semantic meaning of the analyzed tabular features. By leveraging existing
ontologies and domain knowledge, SeFNet opens up new opportunities for sharing
insights between diverse predictive tasks. One such opportunity is the Dataset
Ontology-based Semantic Similarity (DOSS) measure, which quantifies the
similarity between datasets using relations across their features. In this
paper, we present an example of SeFNet prepared for a collection of predictive
tasks in healthcare, with the features' relations derived from the SNOMED-CT
ontology. The proposed SeFNet framework and the accompanying DOSS measure
address the issue of limited contextual information in tabular datasets. By
incorporating domain knowledge and establishing semantic relations between
features, we enhance the potential for meta-learning and enable valuable
insights to be shared across different predictive tasks. | {
"abstract": "Machine learning applications cover a wide range of predictive tasks in which\ntabular datasets play a significant role. However, although they often address\nsimilar problems, tabular datasets are typically treated as standalone tasks.\nThe possibilities of using previously solved problems are limited due to the\nlack of structured contextual information about their features and the lack of\nunderstanding of the relations between them. To overcome this limitation, we\npropose a new approach called Semantic Feature Net (SeFNet), capturing the\nsemantic meaning of the analyzed tabular features. By leveraging existing\nontologies and domain knowledge, SeFNet opens up new opportunities for sharing\ninsights between diverse predictive tasks. One such opportunity is the Dataset\nOntology-based Semantic Similarity (DOSS) measure, which quantifies the\nsimilarity between datasets using relations across their features. In this\npaper, we present an example of SeFNet prepared for a collection of predictive\ntasks in healthcare, with the features' relations derived from the SNOMED-CT\nontology. The proposed SeFNet framework and the accompanying DOSS measure\naddress the issue of limited contextual information in tabular datasets. By\nincorporating domain knowledge and establishing semantic relations between\nfeatures, we enhance the potential for meta-learning and enable valuable\ninsights to be shared across different predictive tasks.",
"title": "SeFNet: Bridging Tabular Datasets with Semantic Feature Nets",
"url": "http://arxiv.org/abs/2306.11636v1"
} | null | null | no_new_dataset | admin | null | false | null | 2866390a-1cbe-4439-9100-a33f47bb810e | null | Validated | 2023-10-04 15:19:51.870316 | {
"text_length": 1527
} | 1no_new_dataset
|
TITLE: A Retinex based GAN Pipeline to Utilize Paired and Unpaired Datasets for Enhancing Low Light Images
ABSTRACT: Low light image enhancement is an important challenge for the development of
robust computer vision algorithms. The machine learning approaches to this have
been either unsupervised, supervised based on paired dataset or supervised
based on unpaired dataset. This paper presents a novel deep learning pipeline
that can learn from both paired and unpaired datasets. Convolution Neural
Networks (CNNs) that are optimized to minimize standard loss, and Generative
Adversarial Networks (GANs) that are optimized to minimize the adversarial loss
are used to achieve different steps of the low light image enhancement process.
Cycle consistency loss and a patched discriminator are utilized to further
improve the performance. The paper also analyses the functionality and the
performance of different components, hidden layers, and the entire pipeline. | {
"abstract": "Low light image enhancement is an important challenge for the development of\nrobust computer vision algorithms. The machine learning approaches to this have\nbeen either unsupervised, supervised based on paired dataset or supervised\nbased on unpaired dataset. This paper presents a novel deep learning pipeline\nthat can learn from both paired and unpaired datasets. Convolution Neural\nNetworks (CNNs) that are optimized to minimize standard loss, and Generative\nAdversarial Networks (GANs) that are optimized to minimize the adversarial loss\nare used to achieve different steps of the low light image enhancement process.\nCycle consistency loss and a patched discriminator are utilized to further\nimprove the performance. The paper also analyses the functionality and the\nperformance of different components, hidden layers, and the entire pipeline.",
"title": "A Retinex based GAN Pipeline to Utilize Paired and Unpaired Datasets for Enhancing Low Light Images",
"url": "http://arxiv.org/abs/2006.15304v2"
} | null | null | no_new_dataset | admin | null | false | null | fa59b76a-fc0c-4d62-afca-b94ff6e467b4 | null | Validated | 2023-10-04 15:19:51.899314 | {
"text_length": 981
} | 1no_new_dataset
|
TITLE: BAN-Cap: A Multi-Purpose English-Bangla Image Descriptions Dataset
ABSTRACT: As computers have become efficient at understanding visual information and
transforming it into a written representation, research interest in tasks like
automatic image captioning has seen a significant leap over the last few years.
While most of the research attention is given to the English language in a
monolingual setting, resource-constrained languages like Bangla remain out of
focus, predominantly due to a lack of standard datasets. Addressing this issue,
we present a new dataset BAN-Cap following the widely used Flickr8k dataset,
where we collect Bangla captions of the images provided by qualified
annotators. Our dataset represents a wider variety of image caption styles
annotated by trained people from different backgrounds. We present a
quantitative and qualitative analysis of the dataset and the baseline
evaluation of the recent models in Bangla image captioning. We investigate the
effect of text augmentation and demonstrate that an adaptive attention-based
model combined with text augmentation using Contextualized Word Replacement
(CWR) outperforms all state-of-the-art models for Bangla image captioning. We
also present this dataset's multipurpose nature, especially on machine
translation for Bangla-English and English-Bangla. This dataset and all the
models will be useful for further research. | {
"abstract": "As computers have become efficient at understanding visual information and\ntransforming it into a written representation, research interest in tasks like\nautomatic image captioning has seen a significant leap over the last few years.\nWhile most of the research attention is given to the English language in a\nmonolingual setting, resource-constrained languages like Bangla remain out of\nfocus, predominantly due to a lack of standard datasets. Addressing this issue,\nwe present a new dataset BAN-Cap following the widely used Flickr8k dataset,\nwhere we collect Bangla captions of the images provided by qualified\nannotators. Our dataset represents a wider variety of image caption styles\nannotated by trained people from different backgrounds. We present a\nquantitative and qualitative analysis of the dataset and the baseline\nevaluation of the recent models in Bangla image captioning. We investigate the\neffect of text augmentation and demonstrate that an adaptive attention-based\nmodel combined with text augmentation using Contextualized Word Replacement\n(CWR) outperforms all state-of-the-art models for Bangla image captioning. We\nalso present this dataset's multipurpose nature, especially on machine\ntranslation for Bangla-English and English-Bangla. This dataset and all the\nmodels will be useful for further research.",
"title": "BAN-Cap: A Multi-Purpose English-Bangla Image Descriptions Dataset",
"url": "http://arxiv.org/abs/2205.14462v1"
} | null | null | new_dataset | admin | null | false | null | cbe96356-b3fc-4dab-8627-2b0c8e5c3e73 | null | Validated | 2023-10-04 15:19:51.886159 | {
"text_length": 1428
} | 0new_dataset
|
TITLE: Multimodal Co-learning: Challenges, Applications with Datasets, Recent Advances and Future Directions
ABSTRACT: Multimodal deep learning systems which employ multiple modalities like text,
image, audio, video, etc., are showing better performance in comparison with
individual modalities (i.e., unimodal) systems. Multimodal machine learning
involves multiple aspects: representation, translation, alignment, fusion, and
co-learning. In the current state of multimodal machine learning, the
assumptions are that all modalities are present, aligned, and noiseless during
training and testing time. However, in real-world tasks, typically, it is
observed that one or more modalities are missing, noisy, lacking annotated
data, have unreliable labels, and are scarce in training or testing and or
both. This challenge is addressed by a learning paradigm called multimodal
co-learning. The modeling of a (resource-poor) modality is aided by exploiting
knowledge from another (resource-rich) modality using transfer of knowledge
between modalities, including their representations and predictive models.
Co-learning being an emerging area, there are no dedicated reviews explicitly
focusing on all challenges addressed by co-learning. To that end, in this work,
we provide a comprehensive survey on the emerging area of multimodal
co-learning that has not been explored in its entirety yet. We review
implementations that overcome one or more co-learning challenges without
explicitly considering them as co-learning challenges. We present the
comprehensive taxonomy of multimodal co-learning based on the challenges
addressed by co-learning and associated implementations. The various techniques
employed to include the latest ones are reviewed along with some of the
applications and datasets. Our final goal is to discuss challenges and
perspectives along with the important ideas and directions for future work that
we hope to be beneficial for the entire research community focusing on this
exciting domain. | {
"abstract": "Multimodal deep learning systems which employ multiple modalities like text,\nimage, audio, video, etc., are showing better performance in comparison with\nindividual modalities (i.e., unimodal) systems. Multimodal machine learning\ninvolves multiple aspects: representation, translation, alignment, fusion, and\nco-learning. In the current state of multimodal machine learning, the\nassumptions are that all modalities are present, aligned, and noiseless during\ntraining and testing time. However, in real-world tasks, typically, it is\nobserved that one or more modalities are missing, noisy, lacking annotated\ndata, have unreliable labels, and are scarce in training or testing and or\nboth. This challenge is addressed by a learning paradigm called multimodal\nco-learning. The modeling of a (resource-poor) modality is aided by exploiting\nknowledge from another (resource-rich) modality using transfer of knowledge\nbetween modalities, including their representations and predictive models.\nCo-learning being an emerging area, there are no dedicated reviews explicitly\nfocusing on all challenges addressed by co-learning. To that end, in this work,\nwe provide a comprehensive survey on the emerging area of multimodal\nco-learning that has not been explored in its entirety yet. We review\nimplementations that overcome one or more co-learning challenges without\nexplicitly considering them as co-learning challenges. We present the\ncomprehensive taxonomy of multimodal co-learning based on the challenges\naddressed by co-learning and associated implementations. The various techniques\nemployed to include the latest ones are reviewed along with some of the\napplications and datasets. Our final goal is to discuss challenges and\nperspectives along with the important ideas and directions for future work that\nwe hope to be beneficial for the entire research community focusing on this\nexciting domain.",
"title": "Multimodal Co-learning: Challenges, Applications with Datasets, Recent Advances and Future Directions",
"url": "http://arxiv.org/abs/2107.13782v3"
} | null | null | no_new_dataset | admin | null | false | null | 198c7cbd-70d8-4594-bf94-561c761c5e25 | null | Validated | 2023-10-04 15:19:51.893405 | {
"text_length": 2031
} | 1no_new_dataset
|
TITLE: Morphset:Augmenting categorical emotion datasets with dimensional affect labels using face morphing
ABSTRACT: Emotion recognition and understanding is a vital component in human-machine
interaction. Dimensional models of affect such as those using valence and
arousal have advantages over traditional categorical ones due to the complexity
of emotional states in humans. However, dimensional emotion annotations are
difficult and expensive to collect, therefore they are not as prevalent in the
affective computing community. To address these issues, we propose a method to
generate synthetic images from existing categorical emotion datasets using face
morphing as well as dimensional labels in the circumplex space with full
control over the resulting sample distribution, while achieving augmentation
factors of at least 20x or more. | {
"abstract": "Emotion recognition and understanding is a vital component in human-machine\ninteraction. Dimensional models of affect such as those using valence and\narousal have advantages over traditional categorical ones due to the complexity\nof emotional states in humans. However, dimensional emotion annotations are\ndifficult and expensive to collect, therefore they are not as prevalent in the\naffective computing community. To address these issues, we propose a method to\ngenerate synthetic images from existing categorical emotion datasets using face\nmorphing as well as dimensional labels in the circumplex space with full\ncontrol over the resulting sample distribution, while achieving augmentation\nfactors of at least 20x or more.",
"title": "Morphset:Augmenting categorical emotion datasets with dimensional affect labels using face morphing",
"url": "http://arxiv.org/abs/2103.02854v2"
} | null | null | no_new_dataset | admin | null | false | null | 17bf625e-84b7-47b0-b890-33228b32a5db | null | Validated | 2023-10-04 15:19:51.895563 | {
"text_length": 860
} | 1no_new_dataset
|
TITLE: Using GAN-based models to sentimental analysis on imbalanced datasets in education domain
ABSTRACT: While the whole world is still struggling with the COVID-19 pandemic, online
learning and home office become more common. Many schools transfer their
courses teaching to the online classroom. Therefore, it is significant to mine
the students' feedback and opinions from their reviews towards studies so that
both schools and teachers can know where they need to improve. This paper
trains machine learning and deep learning models using both balanced and
imbalanced datasets for sentiment classification. Two SOTA category-aware text
generation GAN models: CatGAN and SentiGAN, are utilized to synthesize text
used to balance the highly imbalanced dataset. Results on three datasets with
different imbalance degree from distinct domains show that when using generated
text to balance the dataset, the F1-score of machine learning and deep learning
model on sentiment classification increases 2.79% ~ 9.28%. Also, the results
indicate that the average growth degree for CR100k is higher than CR23k, the
average growth degree for deep learning is more increased than machine learning
algorithms, and the average growth degree for more complex deep learning models
is more increased than simpler deep learning models in experiments. | {
"abstract": "While the whole world is still struggling with the COVID-19 pandemic, online\nlearning and home office become more common. Many schools transfer their\ncourses teaching to the online classroom. Therefore, it is significant to mine\nthe students' feedback and opinions from their reviews towards studies so that\nboth schools and teachers can know where they need to improve. This paper\ntrains machine learning and deep learning models using both balanced and\nimbalanced datasets for sentiment classification. Two SOTA category-aware text\ngeneration GAN models: CatGAN and SentiGAN, are utilized to synthesize text\nused to balance the highly imbalanced dataset. Results on three datasets with\ndifferent imbalance degree from distinct domains show that when using generated\ntext to balance the dataset, the F1-score of machine learning and deep learning\nmodel on sentiment classification increases 2.79% ~ 9.28%. Also, the results\nindicate that the average growth degree for CR100k is higher than CR23k, the\naverage growth degree for deep learning is more increased than machine learning\nalgorithms, and the average growth degree for more complex deep learning models\nis more increased than simpler deep learning models in experiments.",
"title": "Using GAN-based models to sentimental analysis on imbalanced datasets in education domain",
"url": "http://arxiv.org/abs/2108.12061v1"
} | null | null | no_new_dataset | admin | null | false | null | 47978b46-eee3-4e0e-ac75-ab9006820848 | null | Validated | 2023-10-04 15:19:51.892613 | {
"text_length": 1353
} | 1no_new_dataset
|
TITLE: Leak Proof PDBBind: A Reorganized Dataset of Protein-Ligand Complexes for More Generalizable Binding Affinity Prediction
ABSTRACT: Many physics-based and machine-learned scoring functions (SFs) used to
predict protein-ligand binding free energies have been trained on the PDBBind
dataset. However, it is controversial as to whether new SFs are actually
improving since the general, refined, and core datasets of PDBBind are
cross-contaminated with proteins and ligands with high similarity, and hence
they may not perform comparably well in binding prediction of new
protein-ligand complexes. In this work we have carefully prepared a cleaned
PDBBind data set of non-covalent binders that are split into training,
validation, and test datasets to control for data leakage. The resulting
leak-proof (LP)-PDBBind data is used to retrain four popular SFs: AutoDock
vina, Random Forest (RF)-Score, InteractionGraphNet (IGN), and DeepDTA, to
better test their capabilities when applied to new protein-ligand complexes. In
particular we have formulated a new independent data set, BDB2020+, by matching
high quality binding free energies from BindingDB with co-crystalized
ligand-protein complexes from the PDB that have been deposited since 2020.
Based on all the benchmark results, the retrained models using LP-PDBBind that
rely on 3D information perform consistently among the best, with IGN especially
being recommended for scoring and ranking applications for new protein-ligand
systems. | {
"abstract": "Many physics-based and machine-learned scoring functions (SFs) used to\npredict protein-ligand binding free energies have been trained on the PDBBind\ndataset. However, it is controversial as to whether new SFs are actually\nimproving since the general, refined, and core datasets of PDBBind are\ncross-contaminated with proteins and ligands with high similarity, and hence\nthey may not perform comparably well in binding prediction of new\nprotein-ligand complexes. In this work we have carefully prepared a cleaned\nPDBBind data set of non-covalent binders that are split into training,\nvalidation, and test datasets to control for data leakage. The resulting\nleak-proof (LP)-PDBBind data is used to retrain four popular SFs: AutoDock\nvina, Random Forest (RF)-Score, InteractionGraphNet (IGN), and DeepDTA, to\nbetter test their capabilities when applied to new protein-ligand complexes. In\nparticular we have formulated a new independent data set, BDB2020+, by matching\nhigh quality binding free energies from BindingDB with co-crystalized\nligand-protein complexes from the PDB that have been deposited since 2020.\nBased on all the benchmark results, the retrained models using LP-PDBBind that\nrely on 3D information perform consistently among the best, with IGN especially\nbeing recommended for scoring and ranking applications for new protein-ligand\nsystems.",
"title": "Leak Proof PDBBind: A Reorganized Dataset of Protein-Ligand Complexes for More Generalizable Binding Affinity Prediction",
"url": "http://arxiv.org/abs/2308.09639v1"
} | null | null | new_dataset | admin | null | false | null | 1d96d147-fcca-431b-95da-384fc4765b5a | null | Validated | 2023-10-04 15:19:51.864084 | {
"text_length": 1511
} | 0new_dataset
|
TITLE: Handling Imbalanced Datasets Through Optimum-Path Forest
ABSTRACT: In the last decade, machine learning-based approaches became capable of
performing a wide range of complex tasks sometimes better than humans,
demanding a fraction of the time. Such an advance is partially due to the
exponential growth in the amount of data available, which makes it possible to
extract trustworthy real-world information from them. However, such data is
generally imbalanced since some phenomena are more likely than others. Such a
behavior yields considerable influence on the machine learning model's
performance since it becomes biased on the more frequent data it receives.
Despite the considerable amount of machine learning methods, a graph-based
approach has attracted considerable notoriety due to the outstanding
performance over many applications, i.e., the Optimum-Path Forest (OPF). In
this paper, we propose three OPF-based strategies to deal with the imbalance
problem: the $\text{O}^2$PF and the OPF-US, which are novel approaches for
oversampling and undersampling, respectively, as well as a hybrid strategy
combining both approaches. The paper also introduces a set of variants
concerning the strategies mentioned above. Results compared against several
state-of-the-art techniques over public and private datasets confirm the
robustness of the proposed approaches. | {
"abstract": "In the last decade, machine learning-based approaches became capable of\nperforming a wide range of complex tasks sometimes better than humans,\ndemanding a fraction of the time. Such an advance is partially due to the\nexponential growth in the amount of data available, which makes it possible to\nextract trustworthy real-world information from them. However, such data is\ngenerally imbalanced since some phenomena are more likely than others. Such a\nbehavior yields considerable influence on the machine learning model's\nperformance since it becomes biased on the more frequent data it receives.\nDespite the considerable amount of machine learning methods, a graph-based\napproach has attracted considerable notoriety due to the outstanding\nperformance over many applications, i.e., the Optimum-Path Forest (OPF). In\nthis paper, we propose three OPF-based strategies to deal with the imbalance\nproblem: the $\\text{O}^2$PF and the OPF-US, which are novel approaches for\noversampling and undersampling, respectively, as well as a hybrid strategy\ncombining both approaches. The paper also introduces a set of variants\nconcerning the strategies mentioned above. Results compared against several\nstate-of-the-art techniques over public and private datasets confirm the\nrobustness of the proposed approaches.",
"title": "Handling Imbalanced Datasets Through Optimum-Path Forest",
"url": "http://arxiv.org/abs/2202.08934v1"
} | null | null | no_new_dataset | admin | null | false | null | 3f211d11-a2ca-43fb-ad9c-8c33b330b1f0 | null | Validated | 2023-10-04 15:19:51.888247 | {
"text_length": 1392
} | 1no_new_dataset
|
TITLE: HaVQA: A Dataset for Visual Question Answering and Multimodal Research in Hausa Language
ABSTRACT: This paper presents HaVQA, the first multimodal dataset for visual
question-answering (VQA) tasks in the Hausa language. The dataset was created
by manually translating 6,022 English question-answer pairs, which are
associated with 1,555 unique images from the Visual Genome dataset. As a
result, the dataset provides 12,044 gold standard English-Hausa parallel
sentences that were translated in a fashion that guarantees their semantic
match with the corresponding visual information. We conducted several baseline
experiments on the dataset, including visual question answering, visual
question elicitation, text-only and multimodal machine translation. | {
"abstract": "This paper presents HaVQA, the first multimodal dataset for visual\nquestion-answering (VQA) tasks in the Hausa language. The dataset was created\nby manually translating 6,022 English question-answer pairs, which are\nassociated with 1,555 unique images from the Visual Genome dataset. As a\nresult, the dataset provides 12,044 gold standard English-Hausa parallel\nsentences that were translated in a fashion that guarantees their semantic\nmatch with the corresponding visual information. We conducted several baseline\nexperiments on the dataset, including visual question answering, visual\nquestion elicitation, text-only and multimodal machine translation.",
"title": "HaVQA: A Dataset for Visual Question Answering and Multimodal Research in Hausa Language",
"url": "http://arxiv.org/abs/2305.17690v1"
} | null | null | new_dataset | admin | null | false | null | 06a05d1f-2a7d-4469-ac9f-d9e19285d2eb | null | Validated | 2023-10-04 15:19:51.876672 | {
"text_length": 778
} | 0new_dataset
|
TITLE: deeplenstronomy: A dataset simulation package for strong gravitational lensing
ABSTRACT: Automated searches for strong gravitational lensing in optical imaging survey
datasets often employ machine learning and deep learning approaches. These
techniques require more example systems to train the algorithms than have
presently been discovered, which creates a need for simulated images as
training dataset supplements. This work introduces and summarizes
deeplenstronomy, an open-source Python package that enables efficient,
large-scale, and reproducible simulation of images of astronomical systems. A
full suite of unit tests, documentation, and example notebooks are available at
https://deepskies.github.io/deeplenstronomy/ . | {
"abstract": "Automated searches for strong gravitational lensing in optical imaging survey\ndatasets often employ machine learning and deep learning approaches. These\ntechniques require more example systems to train the algorithms than have\npresently been discovered, which creates a need for simulated images as\ntraining dataset supplements. This work introduces and summarizes\ndeeplenstronomy, an open-source Python package that enables efficient,\nlarge-scale, and reproducible simulation of images of astronomical systems. A\nfull suite of unit tests, documentation, and example notebooks are available at\nhttps://deepskies.github.io/deeplenstronomy/ .",
"title": "deeplenstronomy: A dataset simulation package for strong gravitational lensing",
"url": "http://arxiv.org/abs/2102.02830v1"
} | null | null | no_new_dataset | admin | null | false | null | 2c6640ae-789a-41b7-a3ef-6583ca5bcf1e | null | Validated | 2023-10-04 15:19:51.895971 | {
"text_length": 753
} | 1no_new_dataset
|
TITLE: Generating tabular datasets under differential privacy
ABSTRACT: Machine Learning (ML) is accelerating progress across fields and industries,
but relies on accessible and high-quality training data. Some of the most
important datasets are found in biomedical and financial domains in the form of
spreadsheets and relational databases. But this tabular data is often sensitive
in nature. Synthetic data generation offers the potential to unlock sensitive
data, but generative models tend to memorise and regurgitate training data,
which undermines the privacy goal. To remedy this, researchers have
incorporated the mathematical framework of Differential Privacy (DP) into the
training process of deep neural networks. But this creates a trade-off between
the quality and privacy of the resulting data. Generative Adversarial Networks
(GANs) are the dominant paradigm for synthesising tabular data under DP, but
suffer from unstable adversarial training and mode collapse, which are
exacerbated by the privacy constraints and challenging tabular data modality.
This work optimises the quality-privacy trade-off of generative models,
producing higher quality tabular datasets with the same privacy guarantees. We
implement novel end-to-end models that leverage attention mechanisms to learn
reversible tabular representations. We also introduce TableDiffusion, the first
differentially-private diffusion model for tabular data synthesis. Our
experiments show that TableDiffusion produces higher-fidelity synthetic
datasets, avoids the mode collapse problem, and achieves state-of-the-art
performance on privatised tabular data synthesis. By implementing
TableDiffusion to predict the added noise, we enabled it to bypass the
challenges of reconstructing mixed-type tabular data. Overall, the diffusion
paradigm proves vastly more data and privacy efficient than the adversarial
paradigm, due to augmented re-use of each data batch and a smoother iterative
training process. | {
"abstract": "Machine Learning (ML) is accelerating progress across fields and industries,\nbut relies on accessible and high-quality training data. Some of the most\nimportant datasets are found in biomedical and financial domains in the form of\nspreadsheets and relational databases. But this tabular data is often sensitive\nin nature. Synthetic data generation offers the potential to unlock sensitive\ndata, but generative models tend to memorise and regurgitate training data,\nwhich undermines the privacy goal. To remedy this, researchers have\nincorporated the mathematical framework of Differential Privacy (DP) into the\ntraining process of deep neural networks. But this creates a trade-off between\nthe quality and privacy of the resulting data. Generative Adversarial Networks\n(GANs) are the dominant paradigm for synthesising tabular data under DP, but\nsuffer from unstable adversarial training and mode collapse, which are\nexacerbated by the privacy constraints and challenging tabular data modality.\nThis work optimises the quality-privacy trade-off of generative models,\nproducing higher quality tabular datasets with the same privacy guarantees. We\nimplement novel end-to-end models that leverage attention mechanisms to learn\nreversible tabular representations. We also introduce TableDiffusion, the first\ndifferentially-private diffusion model for tabular data synthesis. Our\nexperiments show that TableDiffusion produces higher-fidelity synthetic\ndatasets, avoids the mode collapse problem, and achieves state-of-the-art\nperformance on privatised tabular data synthesis. By implementing\nTableDiffusion to predict the added noise, we enabled it to bypass the\nchallenges of reconstructing mixed-type tabular data. Overall, the diffusion\nparadigm proves vastly more data and privacy efficient than the adversarial\nparadigm, due to augmented re-use of each data batch and a smoother iterative\ntraining process.",
"title": "Generating tabular datasets under differential privacy",
"url": "http://arxiv.org/abs/2308.14784v1"
} | null | null | no_new_dataset | admin | null | false | null | 5ebf97a1-1c3a-49ed-bbb3-2785bc446008 | null | Validated | 2023-10-04 15:19:51.863961 | {
"text_length": 1995
} | 1no_new_dataset
|
TITLE: MosMedData: Chest CT Scans With COVID-19 Related Findings Dataset
ABSTRACT: This dataset contains anonymised human lung computed tomography (CT) scans
with COVID-19 related findings, as well as without such findings. A small
subset of studies has been annotated with binary pixel masks depicting regions
of interests (ground-glass opacifications and consolidations). CT scans were
obtained between 1st of March, 2020 and 25th of April, 2020, and provided by
municipal hospitals in Moscow, Russia. Permanent link:
https://mosmed.ai/datasets/covid19_1110. This dataset is licensed under a
Creative Commons Attribution-NonCommercial-NoDerivs 3.0 Unported (CC BY-NC-ND
3.0) License. Key words: artificial intelligence, COVID-19, machine learning,
dataset, CT, chest, imaging | {
"abstract": "This dataset contains anonymised human lung computed tomography (CT) scans\nwith COVID-19 related findings, as well as without such findings. A small\nsubset of studies has been annotated with binary pixel masks depicting regions\nof interests (ground-glass opacifications and consolidations). CT scans were\nobtained between 1st of March, 2020 and 25th of April, 2020, and provided by\nmunicipal hospitals in Moscow, Russia. Permanent link:\nhttps://mosmed.ai/datasets/covid19_1110. This dataset is licensed under a\nCreative Commons Attribution-NonCommercial-NoDerivs 3.0 Unported (CC BY-NC-ND\n3.0) License. Key words: artificial intelligence, COVID-19, machine learning,\ndataset, CT, chest, imaging",
"title": "MosMedData: Chest CT Scans With COVID-19 Related Findings Dataset",
"url": "http://arxiv.org/abs/2005.06465v1"
} | null | null | new_dataset | admin | null | false | null | c821403a-fee7-4a79-8b67-052f04ae25cd | null | Validated | 2023-10-04 15:19:51.899974 | {
"text_length": 794
} | 0new_dataset
|
TITLE: AirfRANS: High Fidelity Computational Fluid Dynamics Dataset for Approximating Reynolds-Averaged Navier-Stokes Solutions
ABSTRACT: Surrogate models are necessary to optimize meaningful quantities in physical
dynamics as their recursive numerical resolutions are often prohibitively
expensive. It is mainly the case for fluid dynamics and the resolution of
Navier-Stokes equations. However, despite the fast-growing field of data-driven
models for physical systems, reference datasets representing real-world
phenomena are lacking. In this work, we develop AirfRANS, a dataset for
studying the two-dimensional incompressible steady-state Reynolds-Averaged
Navier-Stokes equations over airfoils at a subsonic regime and for different
angles of attacks. We also introduce metrics on the stress forces at the
surface of geometries and visualization of boundary layers to assess the
capabilities of models to accurately predict the meaningful information of the
problem. Finally, we propose deep learning baselines on four machine learning
tasks to study AirfRANS under different constraints for generalization
considerations: big and scarce data regime, Reynolds number, and angle of
attack extrapolation. | {
"abstract": "Surrogate models are necessary to optimize meaningful quantities in physical\ndynamics as their recursive numerical resolutions are often prohibitively\nexpensive. It is mainly the case for fluid dynamics and the resolution of\nNavier-Stokes equations. However, despite the fast-growing field of data-driven\nmodels for physical systems, reference datasets representing real-world\nphenomena are lacking. In this work, we develop AirfRANS, a dataset for\nstudying the two-dimensional incompressible steady-state Reynolds-Averaged\nNavier-Stokes equations over airfoils at a subsonic regime and for different\nangles of attacks. We also introduce metrics on the stress forces at the\nsurface of geometries and visualization of boundary layers to assess the\ncapabilities of models to accurately predict the meaningful information of the\nproblem. Finally, we propose deep learning baselines on four machine learning\ntasks to study AirfRANS under different constraints for generalization\nconsiderations: big and scarce data regime, Reynolds number, and angle of\nattack extrapolation.",
"title": "AirfRANS: High Fidelity Computational Fluid Dynamics Dataset for Approximating Reynolds-Averaged Navier-Stokes Solutions",
"url": "http://arxiv.org/abs/2212.07564v3"
} | null | null | new_dataset | admin | null | false | null | 73c6cab4-855d-4d54-91f6-374c718aa4ca | null | Validated | 2023-10-04 15:19:51.882191 | {
"text_length": 1225
} | 0new_dataset
|
TITLE: Multi-Source Domain Adaptation meets Dataset Distillation through Dataset Dictionary Learning
ABSTRACT: In this paper, we consider the intersection of two problems in machine
learning: Multi-Source Domain Adaptation (MSDA) and Dataset Distillation (DD).
On the one hand, the first considers adapting multiple heterogeneous labeled
source domains to an unlabeled target domain. On the other hand, the second
attacks the problem of synthesizing a small summary containing all the
information about the datasets. We thus consider a new problem called MSDA-DD.
To solve it, we adapt previous works in the MSDA literature, such as
Wasserstein Barycenter Transport and Dataset Dictionary Learning, as well as DD
method Distribution Matching. We thoroughly experiment with this novel problem
on four benchmarks (Caltech-Office 10, Tennessee-Eastman Process, Continuous
Stirred Tank Reactor, and Case Western Reserve University), where we show that,
even with as little as 1 sample per class, one achieves state-of-the-art
adaptation performance. | {
"abstract": "In this paper, we consider the intersection of two problems in machine\nlearning: Multi-Source Domain Adaptation (MSDA) and Dataset Distillation (DD).\nOn the one hand, the first considers adapting multiple heterogeneous labeled\nsource domains to an unlabeled target domain. On the other hand, the second\nattacks the problem of synthesizing a small summary containing all the\ninformation about the datasets. We thus consider a new problem called MSDA-DD.\nTo solve it, we adapt previous works in the MSDA literature, such as\nWasserstein Barycenter Transport and Dataset Dictionary Learning, as well as DD\nmethod Distribution Matching. We thoroughly experiment with this novel problem\non four benchmarks (Caltech-Office 10, Tennessee-Eastman Process, Continuous\nStirred Tank Reactor, and Case Western Reserve University), where we show that,\neven with as little as 1 sample per class, one achieves state-of-the-art\nadaptation performance.",
"title": "Multi-Source Domain Adaptation meets Dataset Distillation through Dataset Dictionary Learning",
"url": "http://arxiv.org/abs/2309.07666v1"
} | null | null | no_new_dataset | admin | null | false | null | 290b44ff-b8d9-416d-b624-09716440ca31 | null | Validated | 2023-10-04 15:19:51.863567 | {
"text_length": 1062
} | 1no_new_dataset
|
TITLE: Human-Machine Collaboration Approaches to Build a Dialogue Dataset for Hate Speech Countering
ABSTRACT: Fighting online hate speech is a challenge that is usually addressed using
Natural Language Processing via automatic detection and removal of hate
content. Besides this approach, counter narratives have emerged as an effective
tool employed by NGOs to respond to online hate on social media platforms. For
this reason, Natural Language Generation is currently being studied as a way to
automatize counter narrative writing. However, the existing resources necessary
to train NLG models are limited to 2-turn interactions (a hate speech and a
counter narrative as response), while in real life, interactions can consist of
multiple turns. In this paper, we present a hybrid approach for dialogical data
collection, which combines the intervention of human expert annotators over
machine generated dialogues obtained using 19 different configurations. The
result of this work is DIALOCONAN, the first dataset comprising over 3000
fictitious multi-turn dialogues between a hater and an NGO operator, covering 6
targets of hate. | {
"abstract": "Fighting online hate speech is a challenge that is usually addressed using\nNatural Language Processing via automatic detection and removal of hate\ncontent. Besides this approach, counter narratives have emerged as an effective\ntool employed by NGOs to respond to online hate on social media platforms. For\nthis reason, Natural Language Generation is currently being studied as a way to\nautomatize counter narrative writing. However, the existing resources necessary\nto train NLG models are limited to 2-turn interactions (a hate speech and a\ncounter narrative as response), while in real life, interactions can consist of\nmultiple turns. In this paper, we present a hybrid approach for dialogical data\ncollection, which combines the intervention of human expert annotators over\nmachine generated dialogues obtained using 19 different configurations. The\nresult of this work is DIALOCONAN, the first dataset comprising over 3000\nfictitious multi-turn dialogues between a hater and an NGO operator, covering 6\ntargets of hate.",
"title": "Human-Machine Collaboration Approaches to Build a Dialogue Dataset for Hate Speech Countering",
"url": "http://arxiv.org/abs/2211.03433v1"
} | null | null | new_dataset | admin | null | false | null | 7846bbb9-9c43-48f8-8277-1d711ec791c7 | null | Validated | 2023-10-04 15:19:51.882988 | {
"text_length": 1152
} | 0new_dataset
|
TITLE: IowaRain: A Statewide Rain Event Dataset Based on Weather Radars and Quantitative Precipitation Estimation
ABSTRACT: Effective environmental planning and management to address climate change
could be achieved through extensive environmental modeling with machine
learning and conventional physical models. In order to develop and improve
these models, practitioners and researchers need comprehensive benchmark
datasets that are prepared and processed with environmental expertise that they
can rely on. This study presents an extensive dataset of rainfall events for
the state of Iowa (2016-2019) acquired from the National Weather Service Next
Generation Weather Radar (NEXRAD) system and processed by a quantitative
precipitation estimation system. The dataset presented in this study could be
used for better disaster monitoring, response and recovery by paving the way
for both predictive and prescriptive modeling. | {
"abstract": "Effective environmental planning and management to address climate change\ncould be achieved through extensive environmental modeling with machine\nlearning and conventional physical models. In order to develop and improve\nthese models, practitioners and researchers need comprehensive benchmark\ndatasets that are prepared and processed with environmental expertise that they\ncan rely on. This study presents an extensive dataset of rainfall events for\nthe state of Iowa (2016-2019) acquired from the National Weather Service Next\nGeneration Weather Radar (NEXRAD) system and processed by a quantitative\nprecipitation estimation system. The dataset presented in this study could be\nused for better disaster monitoring, response and recovery by paving the way\nfor both predictive and prescriptive modeling.",
"title": "IowaRain: A Statewide Rain Event Dataset Based on Weather Radars and Quantitative Precipitation Estimation",
"url": "http://arxiv.org/abs/2107.03432v1"
} | null | null | new_dataset | admin | null | false | null | 3ac5e2b0-1c52-4897-b441-85bb9947955e | null | Validated | 2023-10-04 15:19:51.893765 | {
"text_length": 944
} | 0new_dataset
|
TITLE: No Fair Lunch: A Causal Perspective on Dataset Bias in Machine Learning for Medical Imaging
ABSTRACT: As machine learning methods gain prominence within clinical decision-making,
addressing fairness concerns becomes increasingly urgent. Despite considerable
work dedicated to detecting and ameliorating algorithmic bias, today's methods
are deficient with potentially harmful consequences. Our causal perspective
sheds new light on algorithmic bias, highlighting how different sources of
dataset bias may appear indistinguishable yet require substantially different
mitigation strategies. We introduce three families of causal bias mechanisms
stemming from disparities in prevalence, presentation, and annotation. Our
causal analysis underscores how current mitigation methods tackle only a narrow
and often unrealistic subset of scenarios. We provide a practical three-step
framework for reasoning about fairness in medical imaging, supporting the
development of safe and equitable AI prediction models. | {
"abstract": "As machine learning methods gain prominence within clinical decision-making,\naddressing fairness concerns becomes increasingly urgent. Despite considerable\nwork dedicated to detecting and ameliorating algorithmic bias, today's methods\nare deficient with potentially harmful consequences. Our causal perspective\nsheds new light on algorithmic bias, highlighting how different sources of\ndataset bias may appear indistinguishable yet require substantially different\nmitigation strategies. We introduce three families of causal bias mechanisms\nstemming from disparities in prevalence, presentation, and annotation. Our\ncausal analysis underscores how current mitigation methods tackle only a narrow\nand often unrealistic subset of scenarios. We provide a practical three-step\nframework for reasoning about fairness in medical imaging, supporting the\ndevelopment of safe and equitable AI prediction models.",
"title": "No Fair Lunch: A Causal Perspective on Dataset Bias in Machine Learning for Medical Imaging",
"url": "http://arxiv.org/abs/2307.16526v1"
} | null | null | no_new_dataset | admin | null | false | null | 57c92a3b-05aa-4b68-9728-64013583aa84 | null | Validated | 2023-10-04 15:19:51.864593 | {
"text_length": 1028
} | 1no_new_dataset
|
TITLE: PILArNet: Public Dataset for Particle Imaging Liquid Argon Detectors in High Energy Physics
ABSTRACT: Rapid advancement of machine learning solutions has often coincided with the
production of a test public data set. Such datasets reduce the largest barrier
to entry for tackling a problem -- procuring data -- while also providing a
benchmark to compare different solutions. Furthermore, large datasets have been
used to train high-performing feature finders which are then used in new
approaches to problems beyond that initially defined. In order to encourage the
rapid development in the analysis of data collected using liquid argon time
projection chambers, a class of particle detectors used in high energy physics
experiments, we have produced the PILArNet, first 2D and 3D open dataset to be
used for a couple of key analysis tasks. The initial dataset presented in this
paper contains 300,000 samples simulated and recorded in three different volume
sizes. The dataset is stored efficiently in sparse 2D and 3D matrix format with
auxiliary information about simulated particles in the volume, and is made
available for public research use. In this paper we describe the dataset,
tasks, and the method used to procure the sample. | {
"abstract": "Rapid advancement of machine learning solutions has often coincided with the\nproduction of a test public data set. Such datasets reduce the largest barrier\nto entry for tackling a problem -- procuring data -- while also providing a\nbenchmark to compare different solutions. Furthermore, large datasets have been\nused to train high-performing feature finders which are then used in new\napproaches to problems beyond that initially defined. In order to encourage the\nrapid development in the analysis of data collected using liquid argon time\nprojection chambers, a class of particle detectors used in high energy physics\nexperiments, we have produced the PILArNet, first 2D and 3D open dataset to be\nused for a couple of key analysis tasks. The initial dataset presented in this\npaper contains 300,000 samples simulated and recorded in three different volume\nsizes. The dataset is stored efficiently in sparse 2D and 3D matrix format with\nauxiliary information about simulated particles in the volume, and is made\navailable for public research use. In this paper we describe the dataset,\ntasks, and the method used to procure the sample.",
"title": "PILArNet: Public Dataset for Particle Imaging Liquid Argon Detectors in High Energy Physics",
"url": "http://arxiv.org/abs/2006.01993v1"
} | null | null | new_dataset | admin | null | false | null | 5d89689a-34bb-4daa-bf15-0f736f7442b2 | null | Validated | 2023-10-04 15:19:51.899759 | {
"text_length": 1262
} | 0new_dataset
|
TITLE: CoAP-DoS: An IoT Network Intrusion Dataset
ABSTRACT: The need for secure Internet of Things (IoT) devices is growing as IoT
devices are becoming more integrated into vital networks. Many systems rely on
these devices to remain available and provide reliable service. Denial of
service attacks against IoT devices are a real threat due to the fact these low
power devices are very susceptible to denial-of-service attacks. Machine
learning enabled network intrusion detection systems are effective at
identifying new threats, but they require a large amount of data to work well.
There are many network traffic data sets but very few that focus on IoT network
traffic. Within the IoT network data sets there is a lack of CoAP denial of
service data. We propose a novel data set covering this gap. We develop a new
data set by collecting network traffic from real CoAP denial of service attacks
and compare the data on multiple different machine learning classifiers. We
show that the data set is effective on many classifiers. | {
"abstract": "The need for secure Internet of Things (IoT) devices is growing as IoT\ndevices are becoming more integrated into vital networks. Many systems rely on\nthese devices to remain available and provide reliable service. Denial of\nservice attacks against IoT devices are a real threat due to the fact these low\npower devices are very susceptible to denial-of-service attacks. Machine\nlearning enabled network intrusion detection systems are effective at\nidentifying new threats, but they require a large amount of data to work well.\nThere are many network traffic data sets but very few that focus on IoT network\ntraffic. Within the IoT network data sets there is a lack of CoAP denial of\nservice data. We propose a novel data set covering this gap. We develop a new\ndata set by collecting network traffic from real CoAP denial of service attacks\nand compare the data on multiple different machine learning classifiers. We\nshow that the data set is effective on many classifiers.",
"title": "CoAP-DoS: An IoT Network Intrusion Dataset",
"url": "http://arxiv.org/abs/2206.14341v1"
} | null | null | new_dataset | admin | null | false | null | 96fb32b7-5a67-43e2-bf98-3001963116fe | null | Validated | 2023-10-04 15:19:51.885533 | {
"text_length": 1049
} | 0new_dataset
|
TITLE: An Evaluation of Persian-English Machine Translation Datasets with Transformers
ABSTRACT: Nowadays, many researchers are focusing their attention on the subject of
machine translation (MT). However, Persian machine translation has remained
unexplored despite a vast amount of research being conducted in languages with
high resources, such as English. Moreover, while a substantial amount of
research has been undertaken in statistical machine translation for some
datasets in Persian, there is currently no standard baseline for
transformer-based text2text models on each corpus. This study collected and
analysed the most popular and valuable parallel corpora, which were used for
Persian-English translation. Furthermore, we fine-tuned and evaluated two
state-of-the-art attention-based seq2seq models on each dataset separately (48
results). We hope this paper will assist researchers in comparing their Persian
to English and vice versa machine translation results to a standard baseline. | {
"abstract": "Nowadays, many researchers are focusing their attention on the subject of\nmachine translation (MT). However, Persian machine translation has remained\nunexplored despite a vast amount of research being conducted in languages with\nhigh resources, such as English. Moreover, while a substantial amount of\nresearch has been undertaken in statistical machine translation for some\ndatasets in Persian, there is currently no standard baseline for\ntransformer-based text2text models on each corpus. This study collected and\nanalysed the most popular and valuable parallel corpora, which were used for\nPersian-English translation. Furthermore, we fine-tuned and evaluated two\nstate-of-the-art attention-based seq2seq models on each dataset separately (48\nresults). We hope this paper will assist researchers in comparing their Persian\nto English and vice versa machine translation results to a standard baseline.",
"title": "An Evaluation of Persian-English Machine Translation Datasets with Transformers",
"url": "http://arxiv.org/abs/2302.00321v1"
} | null | null | no_new_dataset | admin | null | false | null | 2956caba-6f3c-43b8-8cc8-0fe41ad51a62 | null | Validated | 2023-10-04 15:19:51.881340 | {
"text_length": 1017
} | 1no_new_dataset
|
TITLE: ConvGeN: Convex space learning improves deep-generative oversampling for tabular imbalanced classification on smaller datasets
ABSTRACT: Data is commonly stored in tabular format. Several fields of research are
prone to small imbalanced tabular data. Supervised Machine Learning on such
data is often difficult due to class imbalance. Synthetic data generation,
i.e., oversampling, is a common remedy used to improve classifier performance.
State-of-the-art linear interpolation approaches, such as LoRAS and ProWRAS can
be used to generate synthetic samples from the convex space of the minority
class to improve classifier performance in such cases. Deep generative networks
are common deep learning approaches for synthetic sample generation, widely
used for synthetic image generation. However, their scope on synthetic tabular
data generation in the context of imbalanced classification is not adequately
explored. In this article, we show that existing deep generative models perform
poorly compared to linear interpolation based approaches for imbalanced
classification problems on smaller tabular datasets. To overcome this, we
propose a deep generative model, ConvGeN that combines the idea of convex space
learning with deep generative models. ConvGeN learns the coefficients for the
convex combinations of the minority class samples, such that the synthetic data
is distinct enough from the majority class. Our benchmarking experiments
demonstrate that our proposed model ConvGeN improves imbalanced classification
on such small datasets, as compared to existing deep generative models, while
being at-par with the existing linear interpolation approaches. Moreover, we
discuss how our model can be used for synthetic tabular data generation in
general, even outside the scope of data imbalance and thus, improves the
overall applicability of convex space learning. | {
"abstract": "Data is commonly stored in tabular format. Several fields of research are\nprone to small imbalanced tabular data. Supervised Machine Learning on such\ndata is often difficult due to class imbalance. Synthetic data generation,\ni.e., oversampling, is a common remedy used to improve classifier performance.\nState-of-the-art linear interpolation approaches, such as LoRAS and ProWRAS can\nbe used to generate synthetic samples from the convex space of the minority\nclass to improve classifier performance in such cases. Deep generative networks\nare common deep learning approaches for synthetic sample generation, widely\nused for synthetic image generation. However, their scope on synthetic tabular\ndata generation in the context of imbalanced classification is not adequately\nexplored. In this article, we show that existing deep generative models perform\npoorly compared to linear interpolation based approaches for imbalanced\nclassification problems on smaller tabular datasets. To overcome this, we\npropose a deep generative model, ConvGeN that combines the idea of convex space\nlearning with deep generative models. ConvGeN learns the coefficients for the\nconvex combinations of the minority class samples, such that the synthetic data\nis distinct enough from the majority class. Our benchmarking experiments\ndemonstrate that our proposed model ConvGeN improves imbalanced classification\non such small datasets, as compared to existing deep generative models, while\nbeing at-par with the existing linear interpolation approaches. Moreover, we\ndiscuss how our model can be used for synthetic tabular data generation in\ngeneral, even outside the scope of data imbalance and thus, improves the\noverall applicability of convex space learning.",
"title": "ConvGeN: Convex space learning improves deep-generative oversampling for tabular imbalanced classification on smaller datasets",
"url": "http://arxiv.org/abs/2206.09812v2"
} | null | null | no_new_dataset | admin | null | false | null | ed417e8d-7419-46e3-b603-a8960b465027 | null | Validated | 2023-10-04 15:19:51.885654 | {
"text_length": 1900
} | 1no_new_dataset
|
TITLE: PMLB v1.0: An open source dataset collection for benchmarking machine learning methods
ABSTRACT: Motivation: Novel machine learning and statistical modeling studies rely on
standardized comparisons to existing methods using well-studied benchmark
datasets. Few tools exist that provide rapid access to many of these datasets
through a standardized, user-friendly interface that integrates well with
popular data science workflows.
Results: This release of PMLB provides the largest collection of diverse,
public benchmark datasets for evaluating new machine learning and data science
methods aggregated in one location. v1.0 introduces a number of critical
improvements developed following discussions with the open-source community.
Availability: PMLB is available at https://github.com/EpistasisLab/pmlb.
Python and R interfaces for PMLB can be installed through the Python Package
Index and Comprehensive R Archive Network, respectively. | {
"abstract": "Motivation: Novel machine learning and statistical modeling studies rely on\nstandardized comparisons to existing methods using well-studied benchmark\ndatasets. Few tools exist that provide rapid access to many of these datasets\nthrough a standardized, user-friendly interface that integrates well with\npopular data science workflows.\n Results: This release of PMLB provides the largest collection of diverse,\npublic benchmark datasets for evaluating new machine learning and data science\nmethods aggregated in one location. v1.0 introduces a number of critical\nimprovements developed following discussions with the open-source community.\n Availability: PMLB is available at https://github.com/EpistasisLab/pmlb.\nPython and R interfaces for PMLB can be installed through the Python Package\nIndex and Comprehensive R Archive Network, respectively.",
"title": "PMLB v1.0: An open source dataset collection for benchmarking machine learning methods",
"url": "http://arxiv.org/abs/2012.00058v3"
} | null | null | new_dataset | admin | null | false | null | 15592bc1-9514-4b5e-83e8-6cc0f09fdbdc | null | Validated | 2023-10-04 15:19:51.896877 | {
"text_length": 968
} | 0new_dataset
|
TITLE: A Novel GDP Prediction Technique based on Transfer Learning using CO2 Emission Dataset
ABSTRACT: In the last 150 years, CO2 concentration in the atmosphere has increased from
280 parts per million to 400 parts per million. This has caused an increase in
the average global temperatures by nearly 0.7 degree centigrade due to the
greenhouse effect. However, the most prosperous states are the highest emitters
of greenhouse gases (specially, CO2). This indicates a strong relationship
between gaseous emissions and the gross domestic product (GDP) of the states.
Such a relationship is highly volatile and nonlinear due to its dependence on
the technological advancements and constantly changing domestic and
international regulatory policies and relations. To analyse such vastly
nonlinear relationships, soft computing techniques has been quite effective as
they can predict a compact solution for multi-variable parameters without any
explicit insight into the internal system functionalities. This paper reports a
novel transfer learning based approach for GDP prediction, which we have termed
as Domain Adapted Transfer Learning for GDP Prediction. In the proposed
approach per capita GDP of different nations is predicted using their CO2
emissions via a model trained on the data of any developed or developing
economy. Results are comparatively presented considering three well-known
regression methods such as Generalized Regression Neural Network, Extreme
Learning Machine and Support Vector Regression. Then the proposed approach is
used to reliably estimate the missing per capita GDP of some of the war-torn
and isolated countries. | {
"abstract": "In the last 150 years, CO2 concentration in the atmosphere has increased from\n280 parts per million to 400 parts per million. This has caused an increase in\nthe average global temperatures by nearly 0.7 degree centigrade due to the\ngreenhouse effect. However, the most prosperous states are the highest emitters\nof greenhouse gases (specially, CO2). This indicates a strong relationship\nbetween gaseous emissions and the gross domestic product (GDP) of the states.\nSuch a relationship is highly volatile and nonlinear due to its dependence on\nthe technological advancements and constantly changing domestic and\ninternational regulatory policies and relations. To analyse such vastly\nnonlinear relationships, soft computing techniques has been quite effective as\nthey can predict a compact solution for multi-variable parameters without any\nexplicit insight into the internal system functionalities. This paper reports a\nnovel transfer learning based approach for GDP prediction, which we have termed\nas Domain Adapted Transfer Learning for GDP Prediction. In the proposed\napproach per capita GDP of different nations is predicted using their CO2\nemissions via a model trained on the data of any developed or developing\neconomy. Results are comparatively presented considering three well-known\nregression methods such as Generalized Regression Neural Network, Extreme\nLearning Machine and Support Vector Regression. Then the proposed approach is\nused to reliably estimate the missing per capita GDP of some of the war-torn\nand isolated countries.",
"title": "A Novel GDP Prediction Technique based on Transfer Learning using CO2 Emission Dataset",
"url": "http://arxiv.org/abs/2005.02856v1"
} | null | null | no_new_dataset | admin | null | false | null | b633774c-ca11-40f1-a293-91e427649dba | null | Validated | 2023-10-04 15:19:51.900217 | {
"text_length": 1666
} | 1no_new_dataset
|
TITLE: KIT MOMA: A Mobile Machines Dataset
ABSTRACT: Mobile machines typically working in a closed site, have a high potential to
utilize autonomous driving technology. However, vigorously thriving development
and innovation are happening mostly in the area of passenger cars. In contrast,
although there are also many research pieces about autonomous driving or
working in mobile machines, a consensus about the SOTA solution is still not
achieved. We believe that the most urgent problem that should be solved is the
absence of a public and challenging visual dataset, which makes the results
from different researches comparable. To address the problem, we publish the
KIT MOMA dataset, including eight classes of commonly used mobile machines,
which can be used as a benchmark to evaluate the SOTA algorithms to detect
mobile construction machines. The view of the gathered images is outside of the
mobile machines since we believe fixed cameras on the ground are more suitable
if all the interesting machines are working in a closed site. Most of the
images in KIT MOMA are in a real scene, whereas some of the images are from the
official website of top construction machine companies. Also, we have evaluated
the performance of YOLO v3 on our dataset, indicating that the SOTA computer
vision algorithms already show an excellent performance for detecting the
mobile machines in a specific working site. Together with the dataset, we also
upload the trained weights, which can be directly used by engineers from the
construction machine industry. The dataset, trained weights, and updates can be
found on our Github. Moreover, the demo can be found on our Youtube. | {
"abstract": "Mobile machines typically working in a closed site, have a high potential to\nutilize autonomous driving technology. However, vigorously thriving development\nand innovation are happening mostly in the area of passenger cars. In contrast,\nalthough there are also many research pieces about autonomous driving or\nworking in mobile machines, a consensus about the SOTA solution is still not\nachieved. We believe that the most urgent problem that should be solved is the\nabsence of a public and challenging visual dataset, which makes the results\nfrom different researches comparable. To address the problem, we publish the\nKIT MOMA dataset, including eight classes of commonly used mobile machines,\nwhich can be used as a benchmark to evaluate the SOTA algorithms to detect\nmobile construction machines. The view of the gathered images is outside of the\nmobile machines since we believe fixed cameras on the ground are more suitable\nif all the interesting machines are working in a closed site. Most of the\nimages in KIT MOMA are in a real scene, whereas some of the images are from the\nofficial website of top construction machine companies. Also, we have evaluated\nthe performance of YOLO v3 on our dataset, indicating that the SOTA computer\nvision algorithms already show an excellent performance for detecting the\nmobile machines in a specific working site. Together with the dataset, we also\nupload the trained weights, which can be directly used by engineers from the\nconstruction machine industry. The dataset, trained weights, and updates can be\nfound on our Github. Moreover, the demo can be found on our Youtube.",
"title": "KIT MOMA: A Mobile Machines Dataset",
"url": "http://arxiv.org/abs/2007.04198v1"
} | null | null | new_dataset | admin | null | false | null | e057ffd6-8604-4ad2-95b9-a1ce6bbb384b | null | Validated | 2023-10-04 15:19:51.899243 | {
"text_length": 1688
} | 0new_dataset
|
TITLE: Chickenpox Cases in Hungary: a Benchmark Dataset for Spatiotemporal Signal Processing with Graph Neural Networks
ABSTRACT: Recurrent graph convolutional neural networks are highly effective machine
learning techniques for spatiotemporal signal processing. Newly proposed graph
neural network architectures are repetitively evaluated on standard tasks such
as traffic or weather forecasting. In this paper, we propose the Chickenpox
Cases in Hungary dataset as a new dataset for comparing graph neural network
architectures. Our time series analysis and forecasting experiments demonstrate
that the Chickenpox Cases in Hungary dataset is adequate for comparing the
predictive performance and forecasting capabilities of novel recurrent graph
neural network architectures. | {
"abstract": "Recurrent graph convolutional neural networks are highly effective machine\nlearning techniques for spatiotemporal signal processing. Newly proposed graph\nneural network architectures are repetitively evaluated on standard tasks such\nas traffic or weather forecasting. In this paper, we propose the Chickenpox\nCases in Hungary dataset as a new dataset for comparing graph neural network\narchitectures. Our time series analysis and forecasting experiments demonstrate\nthat the Chickenpox Cases in Hungary dataset is adequate for comparing the\npredictive performance and forecasting capabilities of novel recurrent graph\nneural network architectures.",
"title": "Chickenpox Cases in Hungary: a Benchmark Dataset for Spatiotemporal Signal Processing with Graph Neural Networks",
"url": "http://arxiv.org/abs/2102.08100v1"
} | null | null | new_dataset | admin | null | false | null | 720976ff-fc64-4f09-a4ac-bc4f30fbb339 | null | Validated | 2023-10-04 15:19:51.895780 | {
"text_length": 794
} | 0new_dataset
|
TITLE: Scalable neural network models and terascale datasets for particle-flow reconstruction
ABSTRACT: We study scalable machine learning models for full event reconstruction in
high-energy electron-positron collisions based on a highly granular detector
simulation. Particle-flow (PF) reconstruction can be formulated as a supervised
learning task using tracks and calorimeter clusters or hits. We compare a graph
neural network and kernel-based transformer and demonstrate that both avoid
quadratic memory allocation and computational cost while achieving realistic PF
reconstruction. We show that hyperparameter tuning on a supercomputer
significantly improves the physics performance of the models. We also
demonstrate that the resulting model is highly portable across hardware
processors, supporting Nvidia, AMD, and Intel Habana cards. Finally, we
demonstrate that the model can be trained on highly granular inputs consisting
of tracks and calorimeter hits, resulting in a competitive physics performance
with the baseline. Datasets and software to reproduce the studies are published
following the findable, accessible, interoperable, and reusable (FAIR)
principles. | {
"abstract": "We study scalable machine learning models for full event reconstruction in\nhigh-energy electron-positron collisions based on a highly granular detector\nsimulation. Particle-flow (PF) reconstruction can be formulated as a supervised\nlearning task using tracks and calorimeter clusters or hits. We compare a graph\nneural network and kernel-based transformer and demonstrate that both avoid\nquadratic memory allocation and computational cost while achieving realistic PF\nreconstruction. We show that hyperparameter tuning on a supercomputer\nsignificantly improves the physics performance of the models. We also\ndemonstrate that the resulting model is highly portable across hardware\nprocessors, supporting Nvidia, AMD, and Intel Habana cards. Finally, we\ndemonstrate that the model can be trained on highly granular inputs consisting\nof tracks and calorimeter hits, resulting in a competitive physics performance\nwith the baseline. Datasets and software to reproduce the studies are published\nfollowing the findable, accessible, interoperable, and reusable (FAIR)\nprinciples.",
"title": "Scalable neural network models and terascale datasets for particle-flow reconstruction",
"url": "http://arxiv.org/abs/2309.06782v1"
} | null | null | no_new_dataset | admin | null | false | null | 5960ed46-fc67-4a6c-89db-fe3462344094 | null | Validated | 2023-10-04 15:19:51.863641 | {
"text_length": 1193
} | 1no_new_dataset
|
TITLE: A Survey on Legal Judgment Prediction: Datasets, Metrics, Models and Challenges
ABSTRACT: Legal judgment prediction (LJP) applies Natural Language Processing (NLP)
techniques to predict judgment results based on fact descriptions
automatically. Recently, large-scale public datasets and advances in NLP
research have led to increasing interest in LJP. Despite a clear gap between
machine and human performance, impressive results have been achieved in various
benchmark datasets. In this paper, to address the current lack of comprehensive
survey of existing LJP tasks, datasets, models and evaluations, (1) we analyze
31 LJP datasets in 6 languages, present their construction process and define a
classification method of LJP with 3 different attributes; (2) we summarize 14
evaluation metrics under four categories for different outputs of LJP tasks;
(3) we review 12 legal-domain pretrained models in 3 languages and highlight 3
major research directions for LJP; (4) we show the state-of-art results for 8
representative datasets from different court cases and discuss the open
challenges. This paper can provide up-to-date and comprehensive reviews to help
readers understand the status of LJP. We hope to facilitate both NLP
researchers and legal professionals for further joint efforts in this problem. | {
"abstract": "Legal judgment prediction (LJP) applies Natural Language Processing (NLP)\ntechniques to predict judgment results based on fact descriptions\nautomatically. Recently, large-scale public datasets and advances in NLP\nresearch have led to increasing interest in LJP. Despite a clear gap between\nmachine and human performance, impressive results have been achieved in various\nbenchmark datasets. In this paper, to address the current lack of comprehensive\nsurvey of existing LJP tasks, datasets, models and evaluations, (1) we analyze\n31 LJP datasets in 6 languages, present their construction process and define a\nclassification method of LJP with 3 different attributes; (2) we summarize 14\nevaluation metrics under four categories for different outputs of LJP tasks;\n(3) we review 12 legal-domain pretrained models in 3 languages and highlight 3\nmajor research directions for LJP; (4) we show the state-of-art results for 8\nrepresentative datasets from different court cases and discuss the open\nchallenges. This paper can provide up-to-date and comprehensive reviews to help\nreaders understand the status of LJP. We hope to facilitate both NLP\nresearchers and legal professionals for further joint efforts in this problem.",
"title": "A Survey on Legal Judgment Prediction: Datasets, Metrics, Models and Challenges",
"url": "http://arxiv.org/abs/2204.04859v1"
} | null | null | no_new_dataset | admin | null | false | null | 09ae0313-c7d2-4440-a6f0-dd3bc741e01a | null | Validated | 2023-10-04 15:19:51.887109 | {
"text_length": 1334
} | 1no_new_dataset
|
TITLE: WikiDes: A Wikipedia-Based Dataset for Generating Short Descriptions from Paragraphs
ABSTRACT: As free online encyclopedias with massive volumes of content, Wikipedia and
Wikidata are key to many Natural Language Processing (NLP) tasks, such as
information retrieval, knowledge base building, machine translation, text
classification, and text summarization. In this paper, we introduce WikiDes, a
novel dataset to generate short descriptions of Wikipedia articles for the
problem of text summarization. The dataset consists of over 80k English samples
on 6987 topics. We set up a two-phase summarization method - description
generation (Phase I) and candidate ranking (Phase II) - as a strong approach
that relies on transfer and contrastive learning. For description generation,
T5 and BART show their superiority compared to other small-scale pre-trained
models. By applying contrastive learning with the diverse input from beam
search, the metric fusion-based ranking models outperform the direct
description generation models significantly up to 22 ROUGE in topic-exclusive
split and topic-independent split. Furthermore, the outcome descriptions in
Phase II are supported by human evaluation in over 45.33% chosen compared to
23.66% in Phase I against the gold descriptions. In the aspect of sentiment
analysis, the generated descriptions cannot effectively capture all sentiment
polarities from paragraphs while doing this task better from the gold
descriptions. The automatic generation of new descriptions reduces the human
efforts in creating them and enriches Wikidata-based knowledge graphs. Our
paper shows a practical impact on Wikipedia and Wikidata since there are
thousands of missing descriptions. Finally, we expect WikiDes to be a useful
dataset for related works in capturing salient information from short
paragraphs. The curated dataset is publicly available at:
https://github.com/declare-lab/WikiDes. | {
"abstract": "As free online encyclopedias with massive volumes of content, Wikipedia and\nWikidata are key to many Natural Language Processing (NLP) tasks, such as\ninformation retrieval, knowledge base building, machine translation, text\nclassification, and text summarization. In this paper, we introduce WikiDes, a\nnovel dataset to generate short descriptions of Wikipedia articles for the\nproblem of text summarization. The dataset consists of over 80k English samples\non 6987 topics. We set up a two-phase summarization method - description\ngeneration (Phase I) and candidate ranking (Phase II) - as a strong approach\nthat relies on transfer and contrastive learning. For description generation,\nT5 and BART show their superiority compared to other small-scale pre-trained\nmodels. By applying contrastive learning with the diverse input from beam\nsearch, the metric fusion-based ranking models outperform the direct\ndescription generation models significantly up to 22 ROUGE in topic-exclusive\nsplit and topic-independent split. Furthermore, the outcome descriptions in\nPhase II are supported by human evaluation in over 45.33% chosen compared to\n23.66% in Phase I against the gold descriptions. In the aspect of sentiment\nanalysis, the generated descriptions cannot effectively capture all sentiment\npolarities from paragraphs while doing this task better from the gold\ndescriptions. The automatic generation of new descriptions reduces the human\nefforts in creating them and enriches Wikidata-based knowledge graphs. Our\npaper shows a practical impact on Wikipedia and Wikidata since there are\nthousands of missing descriptions. Finally, we expect WikiDes to be a useful\ndataset for related works in capturing salient information from short\nparagraphs. The curated dataset is publicly available at:\nhttps://github.com/declare-lab/WikiDes.",
"title": "WikiDes: A Wikipedia-Based Dataset for Generating Short Descriptions from Paragraphs",
"url": "http://arxiv.org/abs/2209.13101v1"
} | null | null | new_dataset | admin | null | false | null | 7b7d27a2-f818-42b4-8518-49a094b5cbad | null | Validated | 2023-10-04 15:19:51.883807 | {
"text_length": 1949
} | 0new_dataset
|
TITLE: Deep Learning-based ECG Classification on Raspberry PI using a Tensorflow Lite Model based on PTB-XL Dataset
ABSTRACT: The number of IoT devices in healthcare is expected to rise sharply due to
increased demand since the COVID-19 pandemic. Deep learning and IoT devices are
being employed to monitor body vitals and automate anomaly detection in
clinical and non-clinical settings. Most of the current technology requires the
transmission of raw data to a remote server, which is not efficient for
resource-constrained IoT devices and embedded systems. Additionally, it is
challenging to develop a machine learning model for ECG classification due to
the lack of an extensive open public database. To an extent, to overcome this
challenge PTB-XL dataset has been used. In this work, we have developed machine
learning models to be deployed on Raspberry Pi. We present an evaluation of our
TensorFlow Model with two classification classes. We also present the
evaluation of the corresponding TensorFlow Lite FlatBuffers to demonstrate
their minimal run-time requirements while maintaining acceptable accuracy. | {
"abstract": "The number of IoT devices in healthcare is expected to rise sharply due to\nincreased demand since the COVID-19 pandemic. Deep learning and IoT devices are\nbeing employed to monitor body vitals and automate anomaly detection in\nclinical and non-clinical settings. Most of the current technology requires the\ntransmission of raw data to a remote server, which is not efficient for\nresource-constrained IoT devices and embedded systems. Additionally, it is\nchallenging to develop a machine learning model for ECG classification due to\nthe lack of an extensive open public database. To an extent, to overcome this\nchallenge PTB-XL dataset has been used. In this work, we have developed machine\nlearning models to be deployed on Raspberry Pi. We present an evaluation of our\nTensorFlow Model with two classification classes. We also present the\nevaluation of the corresponding TensorFlow Lite FlatBuffers to demonstrate\ntheir minimal run-time requirements while maintaining acceptable accuracy.",
"title": "Deep Learning-based ECG Classification on Raspberry PI using a Tensorflow Lite Model based on PTB-XL Dataset",
"url": "http://arxiv.org/abs/2209.00989v1"
} | null | null | no_new_dataset | admin | null | false | null | bf432be5-a787-4967-ad96-435304af3be2 | null | Validated | 2023-10-04 15:19:51.884536 | {
"text_length": 1132
} | 1no_new_dataset
|
TITLE: The GINCO Training Dataset for Web Genre Identification of Documents Out in the Wild
ABSTRACT: This paper presents a new training dataset for automatic genre identification
GINCO, which is based on 1,125 crawled Slovenian web documents that consist of
650 thousand words. Each document was manually annotated for genre with a new
annotation schema that builds upon existing schemata, having primarily clarity
of labels and inter-annotator agreement in mind. The dataset consists of
various challenges related to web-based data, such as machine translated
content, encoding errors, multiple contents presented in one document etc.,
enabling evaluation of classifiers in realistic conditions. The initial machine
learning experiments on the dataset show that (1) pre-Transformer models are
drastically less able to model the phenomena, with macro F1 metrics ranging
around 0.22, while Transformer-based models achieve scores of around 0.58, and
(2) multilingual Transformer models work as well on the task as the monolingual
models that were previously proven to be superior to multilingual models on
standard NLP tasks. | {
"abstract": "This paper presents a new training dataset for automatic genre identification\nGINCO, which is based on 1,125 crawled Slovenian web documents that consist of\n650 thousand words. Each document was manually annotated for genre with a new\nannotation schema that builds upon existing schemata, having primarily clarity\nof labels and inter-annotator agreement in mind. The dataset consists of\nvarious challenges related to web-based data, such as machine translated\ncontent, encoding errors, multiple contents presented in one document etc.,\nenabling evaluation of classifiers in realistic conditions. The initial machine\nlearning experiments on the dataset show that (1) pre-Transformer models are\ndrastically less able to model the phenomena, with macro F1 metrics ranging\naround 0.22, while Transformer-based models achieve scores of around 0.58, and\n(2) multilingual Transformer models work as well on the task as the monolingual\nmodels that were previously proven to be superior to multilingual models on\nstandard NLP tasks.",
"title": "The GINCO Training Dataset for Web Genre Identification of Documents Out in the Wild",
"url": "http://arxiv.org/abs/2201.03857v1"
} | null | null | new_dataset | admin | null | false | null | 5517d078-42b7-48ae-b451-184a9188bf62 | null | Validated | 2023-10-04 15:19:51.888915 | {
"text_length": 1142
} | 0new_dataset
|
TITLE: Whitening and second order optimization both make information in the dataset unusable during training, and can reduce or prevent generalization
ABSTRACT: Machine learning is predicated on the concept of generalization: a model
achieving low error on a sufficiently large training set should also perform
well on novel samples from the same distribution. We show that both data
whitening and second order optimization can harm or entirely prevent
generalization. In general, model training harnesses information contained in
the sample-sample second moment matrix of a dataset. For a general class of
models, namely models with a fully connected first layer, we prove that the
information contained in this matrix is the only information which can be used
to generalize. Models trained using whitened data, or with certain second order
optimization schemes, have less access to this information, resulting in
reduced or nonexistent generalization ability. We experimentally verify these
predictions for several architectures, and further demonstrate that
generalization continues to be harmed even when theoretical requirements are
relaxed. However, we also show experimentally that regularized second order
optimization can provide a practical tradeoff, where training is accelerated
but less information is lost, and generalization can in some circumstances even
improve. | {
"abstract": "Machine learning is predicated on the concept of generalization: a model\nachieving low error on a sufficiently large training set should also perform\nwell on novel samples from the same distribution. We show that both data\nwhitening and second order optimization can harm or entirely prevent\ngeneralization. In general, model training harnesses information contained in\nthe sample-sample second moment matrix of a dataset. For a general class of\nmodels, namely models with a fully connected first layer, we prove that the\ninformation contained in this matrix is the only information which can be used\nto generalize. Models trained using whitened data, or with certain second order\noptimization schemes, have less access to this information, resulting in\nreduced or nonexistent generalization ability. We experimentally verify these\npredictions for several architectures, and further demonstrate that\ngeneralization continues to be harmed even when theoretical requirements are\nrelaxed. However, we also show experimentally that regularized second order\noptimization can provide a practical tradeoff, where training is accelerated\nbut less information is lost, and generalization can in some circumstances even\nimprove.",
"title": "Whitening and second order optimization both make information in the dataset unusable during training, and can reduce or prevent generalization",
"url": "http://arxiv.org/abs/2008.07545v4"
} | null | null | no_new_dataset | admin | null | false | null | 04dd94a0-eb29-4a10-8aa0-c4a83e3f33ef | null | Validated | 2023-10-04 15:19:51.898705 | {
"text_length": 1396
} | 1no_new_dataset
|
TITLE: ComPile: A Large IR Dataset from Production Sources
ABSTRACT: Code is increasingly becoming a core data modality of modern machine learning
research impacting not only the way we write code with conversational agents
like OpenAI's ChatGPT, Google's Bard, or Anthropic's Claude, the way we
translate code from one language into another, but also the compiler
infrastructure underlying the language. While modeling approaches may vary and
representations differ, the targeted tasks often remain the same within the
individual classes of models. Relying solely on the ability of modern models to
extract information from unstructured code does not take advantage of 70 years
of programming language and compiler development by not utilizing the structure
inherent to programs in the data collection. This detracts from the performance
of models working over a tokenized representation of input code and precludes
the use of these models in the compiler itself. To work towards the first
intermediate representation (IR) based models, we fully utilize the LLVM
compiler infrastructure, shared by a number of languages, to generate a 182B
token dataset of LLVM IR. We generated this dataset from programming languages
built on the shared LLVM infrastructure, including Rust, Swift, Julia, and
C/C++, by hooking into LLVM code generation either through the language's
package manager or the compiler directly to extract the dataset of intermediate
representations from production grade programs. Statistical analysis proves the
utility of our dataset not only for large language model training, but also for
the introspection into the code generation process itself with the dataset
showing great promise for machine-learned compiler components. | {
"abstract": "Code is increasingly becoming a core data modality of modern machine learning\nresearch impacting not only the way we write code with conversational agents\nlike OpenAI's ChatGPT, Google's Bard, or Anthropic's Claude, the way we\ntranslate code from one language into another, but also the compiler\ninfrastructure underlying the language. While modeling approaches may vary and\nrepresentations differ, the targeted tasks often remain the same within the\nindividual classes of models. Relying solely on the ability of modern models to\nextract information from unstructured code does not take advantage of 70 years\nof programming language and compiler development by not utilizing the structure\ninherent to programs in the data collection. This detracts from the performance\nof models working over a tokenized representation of input code and precludes\nthe use of these models in the compiler itself. To work towards the first\nintermediate representation (IR) based models, we fully utilize the LLVM\ncompiler infrastructure, shared by a number of languages, to generate a 182B\ntoken dataset of LLVM IR. We generated this dataset from programming languages\nbuilt on the shared LLVM infrastructure, including Rust, Swift, Julia, and\nC/C++, by hooking into LLVM code generation either through the language's\npackage manager or the compiler directly to extract the dataset of intermediate\nrepresentations from production grade programs. Statistical analysis proves the\nutility of our dataset not only for large language model training, but also for\nthe introspection into the code generation process itself with the dataset\nshowing great promise for machine-learned compiler components.",
"title": "ComPile: A Large IR Dataset from Production Sources",
"url": "http://arxiv.org/abs/2309.15432v1"
} | null | null | new_dataset | admin | null | false | null | e7ba666f-2f7c-4b69-8853-6cb6504c9772 | null | Validated | 2023-10-04 15:19:51.863231 | {
"text_length": 1763
} | 0new_dataset
|
TITLE: Shifts 2.0: Extending The Dataset of Real Distributional Shifts
ABSTRACT: Distributional shift, or the mismatch between training and deployment data,
is a significant obstacle to the usage of machine learning in high-stakes
industrial applications, such as autonomous driving and medicine. This creates
a need to be able to assess how robustly ML models generalize as well as the
quality of their uncertainty estimates. Standard ML baseline datasets do not
allow these properties to be assessed, as the training, validation and test
data are often identically distributed. Recently, a range of dedicated
benchmarks have appeared, featuring both distributionally matched and shifted
data. Among these benchmarks, the Shifts dataset stands out in terms of the
diversity of tasks as well as the data modalities it features. While most of
the benchmarks are heavily dominated by 2D image classification tasks, Shifts
contains tabular weather forecasting, machine translation, and vehicle motion
prediction tasks. This enables the robustness properties of models to be
assessed on a diverse set of industrial-scale tasks and either universal or
directly applicable task-specific conclusions to be reached. In this paper, we
extend the Shifts Dataset with two datasets sourced from industrial, high-risk
applications of high societal importance. Specifically, we consider the tasks
of segmentation of white matter Multiple Sclerosis lesions in 3D magnetic
resonance brain images and the estimation of power consumption in marine cargo
vessels. Both tasks feature ubiquitous distributional shifts and a strict
safety requirement due to the high cost of errors. These new datasets will
allow researchers to further explore robust generalization and uncertainty
estimation in new situations. In this work, we provide a description of the
dataset and baseline results for both tasks. | {
"abstract": "Distributional shift, or the mismatch between training and deployment data,\nis a significant obstacle to the usage of machine learning in high-stakes\nindustrial applications, such as autonomous driving and medicine. This creates\na need to be able to assess how robustly ML models generalize as well as the\nquality of their uncertainty estimates. Standard ML baseline datasets do not\nallow these properties to be assessed, as the training, validation and test\ndata are often identically distributed. Recently, a range of dedicated\nbenchmarks have appeared, featuring both distributionally matched and shifted\ndata. Among these benchmarks, the Shifts dataset stands out in terms of the\ndiversity of tasks as well as the data modalities it features. While most of\nthe benchmarks are heavily dominated by 2D image classification tasks, Shifts\ncontains tabular weather forecasting, machine translation, and vehicle motion\nprediction tasks. This enables the robustness properties of models to be\nassessed on a diverse set of industrial-scale tasks and either universal or\ndirectly applicable task-specific conclusions to be reached. In this paper, we\nextend the Shifts Dataset with two datasets sourced from industrial, high-risk\napplications of high societal importance. Specifically, we consider the tasks\nof segmentation of white matter Multiple Sclerosis lesions in 3D magnetic\nresonance brain images and the estimation of power consumption in marine cargo\nvessels. Both tasks feature ubiquitous distributional shifts and a strict\nsafety requirement due to the high cost of errors. These new datasets will\nallow researchers to further explore robust generalization and uncertainty\nestimation in new situations. In this work, we provide a description of the\ndataset and baseline results for both tasks.",
"title": "Shifts 2.0: Extending The Dataset of Real Distributional Shifts",
"url": "http://arxiv.org/abs/2206.15407v2"
} | null | null | new_dataset | admin | null | false | null | bb2169db-72b3-4ce2-85ae-d66da1e2fd03 | null | Validated | 2023-10-04 15:19:51.885462 | {
"text_length": 1897
} | 0new_dataset
|
TITLE: MNISQ: A Large-Scale Quantum Circuit Dataset for Machine Learning on/for Quantum Computers in the NISQ era
ABSTRACT: We introduce the first large-scale dataset, MNISQ, for both the Quantum and
the Classical Machine Learning community during the Noisy Intermediate-Scale
Quantum era. MNISQ consists of 4,950,000 data points organized in 9
subdatasets. Building our dataset from the quantum encoding of classical
information (e.g., MNIST dataset), we deliver a dataset in a dual form: in
quantum form, as circuits, and in classical form, as quantum circuit
descriptions (quantum programming language, QASM). In fact, also the Machine
Learning research related to quantum computers undertakes a dual challenge:
enhancing machine learning exploiting the power of quantum computers, while
also leveraging state-of-the-art classical machine learning methodologies to
help the advancement of quantum computing. Therefore, we perform circuit
classification on our dataset, tackling the task with both quantum and
classical models. In the quantum endeavor, we test our circuit dataset with
Quantum Kernel methods, and we show excellent results up to $97\%$ accuracy. In
the classical world, the underlying quantum mechanical structures within the
quantum circuit data are not trivial. Nevertheless, we test our dataset on
three classical models: Structured State Space sequence model (S4), Transformer
and LSTM. In particular, the S4 model applied on the tokenized QASM sequences
reaches an impressive $77\%$ accuracy. These findings illustrate that quantum
circuit-related datasets are likely to be quantum advantageous, but also that
state-of-the-art machine learning methodologies can competently classify and
recognize quantum circuits. We finally entrust the quantum and classical
machine learning community the fundamental challenge to build more
quantum-classical datasets like ours and to build future benchmarks from our
experiments. The dataset is accessible on GitHub and its circuits are easily
run in qulacs or qiskit. | {
"abstract": "We introduce the first large-scale dataset, MNISQ, for both the Quantum and\nthe Classical Machine Learning community during the Noisy Intermediate-Scale\nQuantum era. MNISQ consists of 4,950,000 data points organized in 9\nsubdatasets. Building our dataset from the quantum encoding of classical\ninformation (e.g., MNIST dataset), we deliver a dataset in a dual form: in\nquantum form, as circuits, and in classical form, as quantum circuit\ndescriptions (quantum programming language, QASM). In fact, also the Machine\nLearning research related to quantum computers undertakes a dual challenge:\nenhancing machine learning exploiting the power of quantum computers, while\nalso leveraging state-of-the-art classical machine learning methodologies to\nhelp the advancement of quantum computing. Therefore, we perform circuit\nclassification on our dataset, tackling the task with both quantum and\nclassical models. In the quantum endeavor, we test our circuit dataset with\nQuantum Kernel methods, and we show excellent results up to $97\\%$ accuracy. In\nthe classical world, the underlying quantum mechanical structures within the\nquantum circuit data are not trivial. Nevertheless, we test our dataset on\nthree classical models: Structured State Space sequence model (S4), Transformer\nand LSTM. In particular, the S4 model applied on the tokenized QASM sequences\nreaches an impressive $77\\%$ accuracy. These findings illustrate that quantum\ncircuit-related datasets are likely to be quantum advantageous, but also that\nstate-of-the-art machine learning methodologies can competently classify and\nrecognize quantum circuits. We finally entrust the quantum and classical\nmachine learning community the fundamental challenge to build more\nquantum-classical datasets like ours and to build future benchmarks from our\nexperiments. The dataset is accessible on GitHub and its circuits are easily\nrun in qulacs or qiskit.",
"title": "MNISQ: A Large-Scale Quantum Circuit Dataset for Machine Learning on/for Quantum Computers in the NISQ era",
"url": "http://arxiv.org/abs/2306.16627v1"
} | null | null | new_dataset | admin | null | false | null | 0b9322bb-fb4b-4408-979c-1f2ac365da9e | null | Validated | 2023-10-04 15:19:51.868773 | {
"text_length": 2046
} | 0new_dataset
|
TITLE: An Instance Selection Algorithm for Big Data in High imbalanced datasets based on LSH
ABSTRACT: Training of Machine Learning (ML) models in real contexts often deals with
big data sets and high-class imbalance samples where the class of interest is
unrepresented (minority class). Practical solutions using classical ML models
address the problem of large data sets using parallel/distributed
implementations of training algorithms, approximate model-based solutions, or
applying instance selection (IS) algorithms to eliminate redundant information.
However, the combined problem of big and high imbalanced datasets has been less
addressed. This work proposes three new methods for IS to be able to deal with
large and imbalanced data sets. The proposed methods use Locality Sensitive
Hashing (LSH) as a base clustering technique, and then three different sampling
methods are applied on top of the clusters (or buckets) generated by LSH. The
algorithms were developed in the Apache Spark framework, guaranteeing their
scalability. The experiments carried out in three different datasets suggest
that the proposed IS methods can improve the performance of a base ML model
between 5% and 19% in terms of the geometric mean. | {
"abstract": "Training of Machine Learning (ML) models in real contexts often deals with\nbig data sets and high-class imbalance samples where the class of interest is\nunrepresented (minority class). Practical solutions using classical ML models\naddress the problem of large data sets using parallel/distributed\nimplementations of training algorithms, approximate model-based solutions, or\napplying instance selection (IS) algorithms to eliminate redundant information.\nHowever, the combined problem of big and high imbalanced datasets has been less\naddressed. This work proposes three new methods for IS to be able to deal with\nlarge and imbalanced data sets. The proposed methods use Locality Sensitive\nHashing (LSH) as a base clustering technique, and then three different sampling\nmethods are applied on top of the clusters (or buckets) generated by LSH. The\nalgorithms were developed in the Apache Spark framework, guaranteeing their\nscalability. The experiments carried out in three different datasets suggest\nthat the proposed IS methods can improve the performance of a base ML model\nbetween 5% and 19% in terms of the geometric mean.",
"title": "An Instance Selection Algorithm for Big Data in High imbalanced datasets based on LSH",
"url": "http://arxiv.org/abs/2210.04310v1"
} | null | null | no_new_dataset | admin | null | false | null | 2587e06f-716b-4f13-9596-bfcabfb1e988 | null | Validated | 2023-10-04 15:19:51.883641 | {
"text_length": 1247
} | 1no_new_dataset
|
TITLE: MangoLeafBD: A Comprehensive Image Dataset to Classify Diseased and Healthy Mango Leaves
ABSTRACT: Agriculture is of one of the few remaining sectors that is yet to receive
proper attention from the machine learning community. The importance of
datasets in the machine learning discipline cannot be overemphasized. The lack
of standard and publicly available datasets related to agriculture impedes
practitioners of this discipline to harness the full benefit of these powerful
computational predictive tools and techniques. To improve this scenario, we
develop, to the best of our knowledge, the first-ever standard, ready-to-use,
and publicly available dataset of mango leaves. The images are collected from
four mango orchards of Bangladesh, one of the top mango-growing countries of
the world. The dataset contains 4000 images of about 1800 distinct leaves
covering seven diseases. Although the dataset is developed using mango leaves
of Bangladesh only, since we deal with diseases that are common across many
countries, this dataset is likely to be applicable to identify mango diseases
in other countries as well, thereby boosting mango yield. This dataset is
expected to draw wide attention from machine learning researchers and
practitioners in the field of automated agriculture. | {
"abstract": "Agriculture is of one of the few remaining sectors that is yet to receive\nproper attention from the machine learning community. The importance of\ndatasets in the machine learning discipline cannot be overemphasized. The lack\nof standard and publicly available datasets related to agriculture impedes\npractitioners of this discipline to harness the full benefit of these powerful\ncomputational predictive tools and techniques. To improve this scenario, we\ndevelop, to the best of our knowledge, the first-ever standard, ready-to-use,\nand publicly available dataset of mango leaves. The images are collected from\nfour mango orchards of Bangladesh, one of the top mango-growing countries of\nthe world. The dataset contains 4000 images of about 1800 distinct leaves\ncovering seven diseases. Although the dataset is developed using mango leaves\nof Bangladesh only, since we deal with diseases that are common across many\ncountries, this dataset is likely to be applicable to identify mango diseases\nin other countries as well, thereby boosting mango yield. This dataset is\nexpected to draw wide attention from machine learning researchers and\npractitioners in the field of automated agriculture.",
"title": "MangoLeafBD: A Comprehensive Image Dataset to Classify Diseased and Healthy Mango Leaves",
"url": "http://arxiv.org/abs/2209.02377v1"
} | null | null | new_dataset | admin | null | false | null | ee1b1522-81b9-44c6-a39f-d3279b89af1c | null | Validated | 2023-10-04 15:19:51.884513 | {
"text_length": 1313
} | 0new_dataset
|
TITLE: Seeing the Unseen: Errors and Bias in Visual Datasets
ABSTRACT: From face recognition in smartphones to automatic routing on self-driving
cars, machine vision algorithms lie in the core of these features. These
systems solve image based tasks by identifying and understanding objects,
subsequently making decisions from these information. However, errors in
datasets are usually induced or even magnified in algorithms, at times
resulting in issues such as recognising black people as gorillas and
misrepresenting ethnicities in search results. This paper tracks the errors in
datasets and their impacts, revealing that a flawed dataset could be a result
of limited categories, incomprehensive sourcing and poor classification. | {
"abstract": "From face recognition in smartphones to automatic routing on self-driving\ncars, machine vision algorithms lie in the core of these features. These\nsystems solve image based tasks by identifying and understanding objects,\nsubsequently making decisions from these information. However, errors in\ndatasets are usually induced or even magnified in algorithms, at times\nresulting in issues such as recognising black people as gorillas and\nmisrepresenting ethnicities in search results. This paper tracks the errors in\ndatasets and their impacts, revealing that a flawed dataset could be a result\nof limited categories, incomprehensive sourcing and poor classification.",
"title": "Seeing the Unseen: Errors and Bias in Visual Datasets",
"url": "http://arxiv.org/abs/2211.01847v1"
} | null | null | no_new_dataset | admin | null | false | null | 122d9848-2c69-4636-994b-dc7dcf9f7f5f | null | Validated | 2023-10-04 15:19:51.883137 | {
"text_length": 751
} | 1no_new_dataset
|
TITLE: Challenges of building medical image datasets for development of deep learning software in stroke
ABSTRACT: Despite the large amount of brain CT data generated in clinical practice, the
availability of CT datasets for deep learning (DL) research is currently
limited. Furthermore, the data can be insufficiently or improperly prepared for
machine learning and thus lead to spurious and irreproducible analyses. This
lack of access to comprehensive and diverse datasets poses a significant
challenge for the development of DL algorithms. In this work, we propose a
complete semi-automatic pipeline to address the challenges of preparing a
clinical brain CT dataset for DL analysis and describe the process of
standardising this heterogeneous dataset. Challenges include handling image
sets with different orientations (axial, sagittal, coronal), different image
types (to view soft tissues or bones) and dimensions, and removing redundant
background. The final pipeline was able to process 5,868/10,659 (45%) CT image
datasets. Reasons for rejection include non-axial data (n=1,920), bone
reformats (n=687), separated skull base/vault images (n=1,226), and
registration failures (n=465). Further format adjustments, including image
cropping, resizing and scaling are also needed for DL processing. Of the axial
scans that were not localisers, bone reformats or split brains, 5,868/6,333
(93%) were accepted, while the remaining 465 failed the registration process.
Appropriate preparation of medical imaging datasets for DL is a costly and
time-intensive process. | {
"abstract": "Despite the large amount of brain CT data generated in clinical practice, the\navailability of CT datasets for deep learning (DL) research is currently\nlimited. Furthermore, the data can be insufficiently or improperly prepared for\nmachine learning and thus lead to spurious and irreproducible analyses. This\nlack of access to comprehensive and diverse datasets poses a significant\nchallenge for the development of DL algorithms. In this work, we propose a\ncomplete semi-automatic pipeline to address the challenges of preparing a\nclinical brain CT dataset for DL analysis and describe the process of\nstandardising this heterogeneous dataset. Challenges include handling image\nsets with different orientations (axial, sagittal, coronal), different image\ntypes (to view soft tissues or bones) and dimensions, and removing redundant\nbackground. The final pipeline was able to process 5,868/10,659 (45%) CT image\ndatasets. Reasons for rejection include non-axial data (n=1,920), bone\nreformats (n=687), separated skull base/vault images (n=1,226), and\nregistration failures (n=465). Further format adjustments, including image\ncropping, resizing and scaling are also needed for DL processing. Of the axial\nscans that were not localisers, bone reformats or split brains, 5,868/6,333\n(93%) were accepted, while the remaining 465 failed the registration process.\nAppropriate preparation of medical imaging datasets for DL is a costly and\ntime-intensive process.",
"title": "Challenges of building medical image datasets for development of deep learning software in stroke",
"url": "http://arxiv.org/abs/2309.15081v1"
} | null | null | no_new_dataset | admin | null | false | null | 51ab76ad-dc61-4303-ac0e-c835ffae77ba | null | Validated | 2023-10-04 15:19:51.863265 | {
"text_length": 1586
} | 1no_new_dataset
|
TITLE: DICES Dataset: Diversity in Conversational AI Evaluation for Safety
ABSTRACT: Machine learning approaches often require training and evaluation datasets
with a clear separation between positive and negative examples. This risks
simplifying and even obscuring the inherent subjectivity present in many tasks.
Preserving such variance in content and diversity in datasets is often
expensive and laborious. This is especially troubling when building safety
datasets for conversational AI systems, as safety is both socially and
culturally situated. To demonstrate this crucial aspect of conversational AI
safety, and to facilitate in-depth model performance analyses, we introduce the
DICES (Diversity In Conversational AI Evaluation for Safety) dataset that
contains fine-grained demographic information about raters, high replication of
ratings per item to ensure statistical power for analyses, and encodes rater
votes as distributions across different demographics to allow for in-depth
explorations of different aggregation strategies. In short, the DICES dataset
enables the observation and measurement of variance, ambiguity, and diversity
in the context of conversational AI safety. We also illustrate how the dataset
offers a basis for establishing metrics to show how raters' ratings can
intersects with demographic categories such as racial/ethnic groups, age
groups, and genders. The goal of DICES is to be used as a shared resource and
benchmark that respects diverse perspectives during safety evaluation of
conversational AI systems. | {
"abstract": "Machine learning approaches often require training and evaluation datasets\nwith a clear separation between positive and negative examples. This risks\nsimplifying and even obscuring the inherent subjectivity present in many tasks.\nPreserving such variance in content and diversity in datasets is often\nexpensive and laborious. This is especially troubling when building safety\ndatasets for conversational AI systems, as safety is both socially and\nculturally situated. To demonstrate this crucial aspect of conversational AI\nsafety, and to facilitate in-depth model performance analyses, we introduce the\nDICES (Diversity In Conversational AI Evaluation for Safety) dataset that\ncontains fine-grained demographic information about raters, high replication of\nratings per item to ensure statistical power for analyses, and encodes rater\nvotes as distributions across different demographics to allow for in-depth\nexplorations of different aggregation strategies. In short, the DICES dataset\nenables the observation and measurement of variance, ambiguity, and diversity\nin the context of conversational AI safety. We also illustrate how the dataset\noffers a basis for establishing metrics to show how raters' ratings can\nintersects with demographic categories such as racial/ethnic groups, age\ngroups, and genders. The goal of DICES is to be used as a shared resource and\nbenchmark that respects diverse perspectives during safety evaluation of\nconversational AI systems.",
"title": "DICES Dataset: Diversity in Conversational AI Evaluation for Safety",
"url": "http://arxiv.org/abs/2306.11247v1"
} | null | null | new_dataset | admin | null | false | null | 6809709f-401b-4800-8e73-ebd62a8e93cb | null | Validated | 2023-10-04 15:19:51.870373 | {
"text_length": 1569
} | 0new_dataset
|
TITLE: MAD: A Scalable Dataset for Language Grounding in Videos from Movie Audio Descriptions
ABSTRACT: The recent and increasing interest in video-language research has driven the
development of large-scale datasets that enable data-intensive machine learning
techniques. In comparison, limited effort has been made at assessing the
fitness of these datasets for the video-language grounding task. Recent works
have begun to discover significant limitations in these datasets, suggesting
that state-of-the-art techniques commonly overfit to hidden dataset biases. In
this work, we present MAD (Movie Audio Descriptions), a novel benchmark that
departs from the paradigm of augmenting existing video datasets with text
annotations and focuses on crawling and aligning available audio descriptions
of mainstream movies. MAD contains over 384,000 natural language sentences
grounded in over 1,200 hours of videos and exhibits a significant reduction in
the currently diagnosed biases for video-language grounding datasets. MAD's
collection strategy enables a novel and more challenging version of
video-language grounding, where short temporal moments (typically seconds long)
must be accurately grounded in diverse long-form videos that can last up to
three hours. We have released MAD's data and baselines code at
https://github.com/Soldelli/MAD. | {
"abstract": "The recent and increasing interest in video-language research has driven the\ndevelopment of large-scale datasets that enable data-intensive machine learning\ntechniques. In comparison, limited effort has been made at assessing the\nfitness of these datasets for the video-language grounding task. Recent works\nhave begun to discover significant limitations in these datasets, suggesting\nthat state-of-the-art techniques commonly overfit to hidden dataset biases. In\nthis work, we present MAD (Movie Audio Descriptions), a novel benchmark that\ndeparts from the paradigm of augmenting existing video datasets with text\nannotations and focuses on crawling and aligning available audio descriptions\nof mainstream movies. MAD contains over 384,000 natural language sentences\ngrounded in over 1,200 hours of videos and exhibits a significant reduction in\nthe currently diagnosed biases for video-language grounding datasets. MAD's\ncollection strategy enables a novel and more challenging version of\nvideo-language grounding, where short temporal moments (typically seconds long)\nmust be accurately grounded in diverse long-form videos that can last up to\nthree hours. We have released MAD's data and baselines code at\nhttps://github.com/Soldelli/MAD.",
"title": "MAD: A Scalable Dataset for Language Grounding in Videos from Movie Audio Descriptions",
"url": "http://arxiv.org/abs/2112.00431v2"
} | null | null | new_dataset | admin | null | false | null | 494c8c53-2524-4b53-a878-e121a46948e2 | null | Validated | 2023-10-04 15:19:51.889500 | {
"text_length": 1363
} | 0new_dataset
|
TITLE: A Survey on Device Behavior Fingerprinting: Data Sources, Techniques, Application Scenarios, and Datasets
ABSTRACT: In the current network-based computing world, where the number of
interconnected devices grows exponentially, their diversity, malfunctions, and
cybersecurity threats are increasing at the same rate. To guarantee the correct
functioning and performance of novel environments such as Smart Cities,
Industry 4.0, or crowdsensing, it is crucial to identify the capabilities of
their devices (e.g., sensors, actuators) and detect potential misbehavior that
may arise due to cyberattacks, system faults, or misconfigurations. With this
goal in mind, a promising research field emerged focusing on creating and
managing fingerprints that model the behavior of both the device actions and
its components. The article at hand studies the recent growth of the device
behavior fingerprinting field in terms of application scenarios, behavioral
sources, and processing and evaluation techniques. First, it performs a
comprehensive review of the device types, behavioral data, and processing and
evaluation techniques used by the most recent and representative research works
dealing with two major scenarios: device identification and device misbehavior
detection. After that, each work is deeply analyzed and compared, emphasizing
its characteristics, advantages, and limitations. This article also provides
researchers with a review of the most relevant characteristics of existing
datasets as most of the novel processing techniques are based on machine
learning and deep learning. Finally, it studies the evolution of these two
scenarios in recent years, providing lessons learned, current trends, and
future research challenges to guide new solutions in the area. | {
"abstract": "In the current network-based computing world, where the number of\ninterconnected devices grows exponentially, their diversity, malfunctions, and\ncybersecurity threats are increasing at the same rate. To guarantee the correct\nfunctioning and performance of novel environments such as Smart Cities,\nIndustry 4.0, or crowdsensing, it is crucial to identify the capabilities of\ntheir devices (e.g., sensors, actuators) and detect potential misbehavior that\nmay arise due to cyberattacks, system faults, or misconfigurations. With this\ngoal in mind, a promising research field emerged focusing on creating and\nmanaging fingerprints that model the behavior of both the device actions and\nits components. The article at hand studies the recent growth of the device\nbehavior fingerprinting field in terms of application scenarios, behavioral\nsources, and processing and evaluation techniques. First, it performs a\ncomprehensive review of the device types, behavioral data, and processing and\nevaluation techniques used by the most recent and representative research works\ndealing with two major scenarios: device identification and device misbehavior\ndetection. After that, each work is deeply analyzed and compared, emphasizing\nits characteristics, advantages, and limitations. This article also provides\nresearchers with a review of the most relevant characteristics of existing\ndatasets as most of the novel processing techniques are based on machine\nlearning and deep learning. Finally, it studies the evolution of these two\nscenarios in recent years, providing lessons learned, current trends, and\nfuture research challenges to guide new solutions in the area.",
"title": "A Survey on Device Behavior Fingerprinting: Data Sources, Techniques, Application Scenarios, and Datasets",
"url": "http://arxiv.org/abs/2008.03343v2"
} | null | null | no_new_dataset | admin | null | false | null | 0027e759-c769-4801-a8cd-a91db24975c6 | null | Validated | 2023-10-04 15:19:51.898850 | {
"text_length": 1797
} | 1no_new_dataset
|
TITLE: Dichotomic Pattern Mining with Applications to Intent Prediction from Semi-Structured Clickstream Datasets
ABSTRACT: We introduce a pattern mining framework that operates on semi-structured
datasets and exploits the dichotomy between outcomes. Our approach takes
advantage of constraint reasoning to find sequential patterns that occur
frequently and exhibit desired properties. This allows the creation of novel
pattern embeddings that are useful for knowledge extraction and predictive
modeling. Finally, we present an application on customer intent prediction from
digital clickstream data. Overall, we show that pattern embeddings play an
integrator role between semi-structured data and machine learning models,
improve the performance of the downstream task and retain interpretability. | {
"abstract": "We introduce a pattern mining framework that operates on semi-structured\ndatasets and exploits the dichotomy between outcomes. Our approach takes\nadvantage of constraint reasoning to find sequential patterns that occur\nfrequently and exhibit desired properties. This allows the creation of novel\npattern embeddings that are useful for knowledge extraction and predictive\nmodeling. Finally, we present an application on customer intent prediction from\ndigital clickstream data. Overall, we show that pattern embeddings play an\nintegrator role between semi-structured data and machine learning models,\nimprove the performance of the downstream task and retain interpretability.",
"title": "Dichotomic Pattern Mining with Applications to Intent Prediction from Semi-Structured Clickstream Datasets",
"url": "http://arxiv.org/abs/2201.09178v1"
} | null | null | no_new_dataset | admin | null | false | null | c9bccfa7-03b6-4ce2-b4f2-7561980aba73 | null | Validated | 2023-10-04 15:19:51.888677 | {
"text_length": 816
} | 1no_new_dataset
|
TITLE: Semi-Supervised Semantic Segmentation in Earth Observation: The MiniFrance Suite, Dataset Analysis and Multi-task Network Study
ABSTRACT: The development of semi-supervised learning techniques is essential to
enhance the generalization capacities of machine learning algorithms. Indeed,
raw image data are abundant while labels are scarce, therefore it is crucial to
leverage unlabeled inputs to build better models. The availability of large
databases have been key for the development of learning algorithms with high
level performance.
Despite the major role of machine learning in Earth Observation to derive
products such as land cover maps, datasets in the field are still limited,
either because of modest surface coverage, lack of variety of scenes or
restricted classes to identify. We introduce a novel large-scale dataset for
semi-supervised semantic segmentation in Earth Observation, the MiniFrance
suite. MiniFrance has several unprecedented properties: it is large-scale,
containing over 2000 very high resolution aerial images, accounting for more
than 200 billions samples (pixels); it is varied, covering 16 conurbations in
France, with various climates, different landscapes, and urban as well as
countryside scenes; and it is challenging, considering land use classes with
high-level semantics. Nevertheless, the most distinctive quality of MiniFrance
is being the only dataset in the field especially designed for semi-supervised
learning: it contains labeled and unlabeled images in its training partition,
which reproduces a life-like scenario. Along with this dataset, we present
tools for data representativeness analysis in terms of appearance similarity
and a thorough study of MiniFrance data, demonstrating that it is suitable for
learning and generalizes well in a semi-supervised setting. Finally, we present
semi-supervised deep architectures based on multi-task learning and the first
experiments on MiniFrance. | {
"abstract": "The development of semi-supervised learning techniques is essential to\nenhance the generalization capacities of machine learning algorithms. Indeed,\nraw image data are abundant while labels are scarce, therefore it is crucial to\nleverage unlabeled inputs to build better models. The availability of large\ndatabases have been key for the development of learning algorithms with high\nlevel performance.\n Despite the major role of machine learning in Earth Observation to derive\nproducts such as land cover maps, datasets in the field are still limited,\neither because of modest surface coverage, lack of variety of scenes or\nrestricted classes to identify. We introduce a novel large-scale dataset for\nsemi-supervised semantic segmentation in Earth Observation, the MiniFrance\nsuite. MiniFrance has several unprecedented properties: it is large-scale,\ncontaining over 2000 very high resolution aerial images, accounting for more\nthan 200 billions samples (pixels); it is varied, covering 16 conurbations in\nFrance, with various climates, different landscapes, and urban as well as\ncountryside scenes; and it is challenging, considering land use classes with\nhigh-level semantics. Nevertheless, the most distinctive quality of MiniFrance\nis being the only dataset in the field especially designed for semi-supervised\nlearning: it contains labeled and unlabeled images in its training partition,\nwhich reproduces a life-like scenario. Along with this dataset, we present\ntools for data representativeness analysis in terms of appearance similarity\nand a thorough study of MiniFrance data, demonstrating that it is suitable for\nlearning and generalizes well in a semi-supervised setting. Finally, we present\nsemi-supervised deep architectures based on multi-task learning and the first\nexperiments on MiniFrance.",
"title": "Semi-Supervised Semantic Segmentation in Earth Observation: The MiniFrance Suite, Dataset Analysis and Multi-task Network Study",
"url": "http://arxiv.org/abs/2010.07830v1"
} | null | null | new_dataset | admin | null | false | null | 27070983-724c-4a1f-b90c-e1d8738d2816 | null | Validated | 2023-10-04 15:19:51.897668 | {
"text_length": 1970
} | 0new_dataset
|
TITLE: AutoTherm: A Dataset and Ablation Study for Thermal Comfort Prediction in Vehicles
ABSTRACT: State recognition in well-known and customizable environments such as
vehicles enables novel insights into users and potentially their intentions.
Besides safety-relevant insights into, for example, fatigue, user
experience-related assessments become increasingly relevant. As thermal comfort
is vital for overall comfort, we introduce a dataset for its prediction in
vehicles incorporating 31 input signals and self-labeled user ratings based on
a 7-point Likert scale (-3 to +3) by 21 subjects. An importance ranking of such
signals indicates higher impact on prediction for signals like ambient
temperature, ambient humidity, radiation temperature, and skin temperature.
Leveraging modern machine learning architectures enables us to not only
automatically recognize human thermal comfort state but also predict future
states. We provide details on how we train a recurrent network-based classifier
and, thus, perform an initial performance benchmark of our proposed thermal
comfort dataset. Ultimately, we compare our collected dataset to publicly
available datasets. | {
"abstract": "State recognition in well-known and customizable environments such as\nvehicles enables novel insights into users and potentially their intentions.\nBesides safety-relevant insights into, for example, fatigue, user\nexperience-related assessments become increasingly relevant. As thermal comfort\nis vital for overall comfort, we introduce a dataset for its prediction in\nvehicles incorporating 31 input signals and self-labeled user ratings based on\na 7-point Likert scale (-3 to +3) by 21 subjects. An importance ranking of such\nsignals indicates higher impact on prediction for signals like ambient\ntemperature, ambient humidity, radiation temperature, and skin temperature.\nLeveraging modern machine learning architectures enables us to not only\nautomatically recognize human thermal comfort state but also predict future\nstates. We provide details on how we train a recurrent network-based classifier\nand, thus, perform an initial performance benchmark of our proposed thermal\ncomfort dataset. Ultimately, we compare our collected dataset to publicly\navailable datasets.",
"title": "AutoTherm: A Dataset and Ablation Study for Thermal Comfort Prediction in Vehicles",
"url": "http://arxiv.org/abs/2211.08257v2"
} | null | null | new_dataset | admin | null | false | null | 20e72ef1-e5f5-4d90-b73e-0de5f43f6030 | null | Validated | 2023-10-04 15:19:51.882771 | {
"text_length": 1188
} | 0new_dataset
|
TITLE: 2DeteCT -- A large 2D expandable, trainable, experimental Computed Tomography dataset for machine learning
ABSTRACT: Recent research in computational imaging largely focuses on developing
machine learning (ML) techniques for image reconstruction, which requires
large-scale training datasets consisting of measurement data and ground-truth
images. However, suitable experimental datasets for X-ray Computed Tomography
(CT) are scarce, and methods are often developed and evaluated only on
simulated data. We fill this gap by providing the community with a versatile,
open 2D fan-beam CT dataset suitable for developing ML techniques for a range
of image reconstruction tasks. To acquire it, we designed a sophisticated,
semi-automatic scan procedure that utilizes a highly-flexible laboratory X-ray
CT setup. A diverse mix of samples with high natural variability in shape and
density was scanned slice-by-slice (5000 slices in total) with high angular and
spatial resolution and three different beam characteristics: A high-fidelity, a
low-dose and a beam-hardening-inflicted mode. In addition, 750
out-of-distribution slices were scanned with sample and beam variations to
accommodate robustness and segmentation tasks. We provide raw projection data,
reference reconstructions and segmentations based on an open-source data
processing pipeline. | {
"abstract": "Recent research in computational imaging largely focuses on developing\nmachine learning (ML) techniques for image reconstruction, which requires\nlarge-scale training datasets consisting of measurement data and ground-truth\nimages. However, suitable experimental datasets for X-ray Computed Tomography\n(CT) are scarce, and methods are often developed and evaluated only on\nsimulated data. We fill this gap by providing the community with a versatile,\nopen 2D fan-beam CT dataset suitable for developing ML techniques for a range\nof image reconstruction tasks. To acquire it, we designed a sophisticated,\nsemi-automatic scan procedure that utilizes a highly-flexible laboratory X-ray\nCT setup. A diverse mix of samples with high natural variability in shape and\ndensity was scanned slice-by-slice (5000 slices in total) with high angular and\nspatial resolution and three different beam characteristics: A high-fidelity, a\nlow-dose and a beam-hardening-inflicted mode. In addition, 750\nout-of-distribution slices were scanned with sample and beam variations to\naccommodate robustness and segmentation tasks. We provide raw projection data,\nreference reconstructions and segmentations based on an open-source data\nprocessing pipeline.",
"title": "2DeteCT -- A large 2D expandable, trainable, experimental Computed Tomography dataset for machine learning",
"url": "http://arxiv.org/abs/2306.05907v1"
} | null | null | new_dataset | admin | null | false | null | 22fd94c7-7839-4009-9a73-65df0b8615f0 | null | Validated | 2023-10-04 15:19:51.874082 | {
"text_length": 1371
} | 0new_dataset
|
TITLE: MIMII DUE: Sound Dataset for Malfunctioning Industrial Machine Investigation and Inspection with Domain Shifts due to Changes in Operational and Environmental Conditions
ABSTRACT: In this paper, we introduce MIMII DUE, a new dataset for malfunctioning
industrial machine investigation and inspection with domain shifts due to
changes in operational and environmental conditions. Conventional methods for
anomalous sound detection face practical challenges because the distribution of
features changes between the training and operational phases (called domain
shift) due to various real-world factors. To check the robustness against
domain shifts, we need a dataset that actually includes domain shifts, but such
a dataset does not exist so far. The new dataset we created consists of the
normal and abnormal operating sounds of five different types of industrial
machines under two different operational/environmental conditions (source
domain and target domain) independent of normal/abnormal, with domain shifts
occurring between the two domains. Experimental results showed significant
performance differences between the source and target domains, indicating that
the dataset contains the domain shifts. These findings demonstrate that the
dataset will be helpful for checking the robustness against domain shifts. The
dataset is a subset of the dataset for DCASE 2021 Challenge Task 2 and freely
available for download at https://zenodo.org/record/4740355 | {
"abstract": "In this paper, we introduce MIMII DUE, a new dataset for malfunctioning\nindustrial machine investigation and inspection with domain shifts due to\nchanges in operational and environmental conditions. Conventional methods for\nanomalous sound detection face practical challenges because the distribution of\nfeatures changes between the training and operational phases (called domain\nshift) due to various real-world factors. To check the robustness against\ndomain shifts, we need a dataset that actually includes domain shifts, but such\na dataset does not exist so far. The new dataset we created consists of the\nnormal and abnormal operating sounds of five different types of industrial\nmachines under two different operational/environmental conditions (source\ndomain and target domain) independent of normal/abnormal, with domain shifts\noccurring between the two domains. Experimental results showed significant\nperformance differences between the source and target domains, indicating that\nthe dataset contains the domain shifts. These findings demonstrate that the\ndataset will be helpful for checking the robustness against domain shifts. The\ndataset is a subset of the dataset for DCASE 2021 Challenge Task 2 and freely\navailable for download at https://zenodo.org/record/4740355",
"title": "MIMII DUE: Sound Dataset for Malfunctioning Industrial Machine Investigation and Inspection with Domain Shifts due to Changes in Operational and Environmental Conditions",
"url": "http://arxiv.org/abs/2105.02702v3"
} | null | null | new_dataset | admin | null | false | null | 75df841f-127a-4a7f-9701-08783316c577 | null | Validated | 2023-10-04 15:19:51.894642 | {
"text_length": 1486
} | 0new_dataset
|
TITLE: Whose Ground Truth? Accounting for Individual and Collective Identities Underlying Dataset Annotation
ABSTRACT: Human annotations play a crucial role in machine learning (ML) research and
development. However, the ethical considerations around the processes and
decisions that go into building ML datasets has not received nearly enough
attention. In this paper, we survey an array of literature that provides
insights into ethical considerations around crowdsourced dataset annotation. We
synthesize these insights, and lay out the challenges in this space along two
layers: (1) who the annotator is, and how the annotators' lived experiences can
impact their annotations, and (2) the relationship between the annotators and
the crowdsourcing platforms and what that relationship affords them. Finally,
we put forth a concrete set of recommendations and considerations for dataset
developers at various stages of the ML data pipeline: task formulation,
selection of annotators, platform and infrastructure choices, dataset analysis
and evaluation, and dataset documentation and release. | {
"abstract": "Human annotations play a crucial role in machine learning (ML) research and\ndevelopment. However, the ethical considerations around the processes and\ndecisions that go into building ML datasets has not received nearly enough\nattention. In this paper, we survey an array of literature that provides\ninsights into ethical considerations around crowdsourced dataset annotation. We\nsynthesize these insights, and lay out the challenges in this space along two\nlayers: (1) who the annotator is, and how the annotators' lived experiences can\nimpact their annotations, and (2) the relationship between the annotators and\nthe crowdsourcing platforms and what that relationship affords them. Finally,\nwe put forth a concrete set of recommendations and considerations for dataset\ndevelopers at various stages of the ML data pipeline: task formulation,\nselection of annotators, platform and infrastructure choices, dataset analysis\nand evaluation, and dataset documentation and release.",
"title": "Whose Ground Truth? Accounting for Individual and Collective Identities Underlying Dataset Annotation",
"url": "http://arxiv.org/abs/2112.04554v1"
} | null | null | no_new_dataset | admin | null | false | null | 177f6692-1afd-4b68-bf91-b0d32cdbf396 | null | Validated | 2023-10-04 15:19:51.889347 | {
"text_length": 1111
} | 1no_new_dataset
|
TITLE: Intrusion Detection Systems Using Support Vector Machines on the KDDCUP'99 and NSL-KDD Datasets: A Comprehensive Survey
ABSTRACT: With the growing rates of cyber-attacks and cyber espionage, the need for
better and more powerful intrusion detection systems (IDS) is even more
warranted nowadays. The basic task of an IDS is to act as the first line of
defense, in detecting attacks on the internet. As intrusion tactics from
intruders become more sophisticated and difficult to detect, researchers have
started to apply novel Machine Learning (ML) techniques to effectively detect
intruders and hence preserve internet users' information and overall trust in
the entire internet network security. Over the last decade, there has been an
explosion of research on intrusion detection techniques based on ML and Deep
Learning (DL) architectures on various cyber security-based datasets such as
the DARPA, KDDCUP'99, NSL-KDD, CAIDA, CTU-13, UNSW-NB15. In this research, we
review contemporary literature and provide a comprehensive survey of different
types of intrusion detection technique that applies Support Vector Machines
(SVMs) algorithms as a classifier. We focus only on studies that have been
evaluated on the two most widely used datasets in cybersecurity namely: the
KDDCUP'99 and the NSL-KDD datasets. We provide a summary of each method,
identifying the role of the SVMs classifier, and all other algorithms involved
in the studies. Furthermore, we present a critical review of each method, in
tabular form, highlighting the performance measures, strengths, and limitations
of each of the methods surveyed. | {
"abstract": "With the growing rates of cyber-attacks and cyber espionage, the need for\nbetter and more powerful intrusion detection systems (IDS) is even more\nwarranted nowadays. The basic task of an IDS is to act as the first line of\ndefense, in detecting attacks on the internet. As intrusion tactics from\nintruders become more sophisticated and difficult to detect, researchers have\nstarted to apply novel Machine Learning (ML) techniques to effectively detect\nintruders and hence preserve internet users' information and overall trust in\nthe entire internet network security. Over the last decade, there has been an\nexplosion of research on intrusion detection techniques based on ML and Deep\nLearning (DL) architectures on various cyber security-based datasets such as\nthe DARPA, KDDCUP'99, NSL-KDD, CAIDA, CTU-13, UNSW-NB15. In this research, we\nreview contemporary literature and provide a comprehensive survey of different\ntypes of intrusion detection technique that applies Support Vector Machines\n(SVMs) algorithms as a classifier. We focus only on studies that have been\nevaluated on the two most widely used datasets in cybersecurity namely: the\nKDDCUP'99 and the NSL-KDD datasets. We provide a summary of each method,\nidentifying the role of the SVMs classifier, and all other algorithms involved\nin the studies. Furthermore, we present a critical review of each method, in\ntabular form, highlighting the performance measures, strengths, and limitations\nof each of the methods surveyed.",
"title": "Intrusion Detection Systems Using Support Vector Machines on the KDDCUP'99 and NSL-KDD Datasets: A Comprehensive Survey",
"url": "http://arxiv.org/abs/2209.05579v1"
} | null | null | no_new_dataset | admin | null | false | null | 0b5e25e8-a4cd-4de2-818b-046552e7461f | null | Validated | 2023-10-04 15:19:51.884054 | {
"text_length": 1640
} | 1no_new_dataset
|
TITLE: Machine Learning Performance Analysis to Predict Stroke Based on Imbalanced Medical Dataset
ABSTRACT: Cerebral stroke, the second most substantial cause of death universally, has
been a primary public health concern over the last few years. With the help of
machine learning techniques, early detection of various stroke alerts is
accessible, which can efficiently prevent or diminish the stroke. Medical
dataset, however, are frequently unbalanced in their class label, with a
tendency to poorly predict minority classes. In this paper, the potential risk
factors for stroke are investigated. Moreover, four distinctive approaches are
applied to improve the classification of the minority class in the imbalanced
stroke dataset, which are the ensemble weight voting classifier, the Synthetic
Minority Over-sampling Technique (SMOTE), Principal Component Analysis with
K-Means Clustering (PCA-Kmeans), Focal Loss with the Deep Neural Network (DNN)
and compare their performance. Through the analysis results, SMOTE and
PCA-Kmeans with DNN-Focal Loss work best for the limited size of a large severe
imbalanced dataset,which is 2-4 times outperform Kaggle work. | {
"abstract": "Cerebral stroke, the second most substantial cause of death universally, has\nbeen a primary public health concern over the last few years. With the help of\nmachine learning techniques, early detection of various stroke alerts is\naccessible, which can efficiently prevent or diminish the stroke. Medical\ndataset, however, are frequently unbalanced in their class label, with a\ntendency to poorly predict minority classes. In this paper, the potential risk\nfactors for stroke are investigated. Moreover, four distinctive approaches are\napplied to improve the classification of the minority class in the imbalanced\nstroke dataset, which are the ensemble weight voting classifier, the Synthetic\nMinority Over-sampling Technique (SMOTE), Principal Component Analysis with\nK-Means Clustering (PCA-Kmeans), Focal Loss with the Deep Neural Network (DNN)\nand compare their performance. Through the analysis results, SMOTE and\nPCA-Kmeans with DNN-Focal Loss work best for the limited size of a large severe\nimbalanced dataset,which is 2-4 times outperform Kaggle work.",
"title": "Machine Learning Performance Analysis to Predict Stroke Based on Imbalanced Medical Dataset",
"url": "http://arxiv.org/abs/2211.07652v1"
} | null | null | no_new_dataset | admin | null | false | null | 935e37df-adc5-49be-835a-24ac558f5c98 | null | Validated | 2023-10-04 15:19:51.882794 | {
"text_length": 1184
} | 1no_new_dataset
|
TITLE: Causal Discovery on the Effect of Antipsychotic Drugs on Delirium Patients in the ICU using Large EHR Dataset
ABSTRACT: Delirium occurs in about 80% cases in the Intensive Care Unit (ICU) and is
associated with a longer hospital stay, increased mortality and other related
issues. Delirium does not have any biomarker-based diagnosis and is commonly
treated with antipsychotic drugs (APD). However, multiple studies have shown
controversy over the efficacy or safety of APD in treating delirium. Since
randomized controlled trials (RCT) are costly and time-expensive, we aim to
approach the research question of the efficacy of APD in the treatment of
delirium using retrospective cohort analysis. We plan to use the Causal
inference framework to look for the underlying causal structure model,
leveraging the availability of large observational data on ICU patients. To
explore safety outcomes associated with APD, we aim to build a causal model for
delirium in the ICU using large observational data sets connecting various
covariates correlated with delirium. We utilized the MIMIC III database, an
extensive electronic health records (EHR) dataset with 53,423 distinct hospital
admissions. Our null hypothesis is: there is no significant difference in
outcomes for delirium patients under different drug-group in the ICU. Through
our exploratory, machine learning based and causal analysis, we had findings
such as: mean length-of-stay and max length-of-stay is higher for patients in
Haloperidol drug group, and haloperidol group has a higher rate of death in a
year compared to other two-groups. Our generated causal model explicitly shows
the functional relationships between different covariates. For future work, we
plan to do time-varying analysis on the dataset. | {
"abstract": "Delirium occurs in about 80% cases in the Intensive Care Unit (ICU) and is\nassociated with a longer hospital stay, increased mortality and other related\nissues. Delirium does not have any biomarker-based diagnosis and is commonly\ntreated with antipsychotic drugs (APD). However, multiple studies have shown\ncontroversy over the efficacy or safety of APD in treating delirium. Since\nrandomized controlled trials (RCT) are costly and time-expensive, we aim to\napproach the research question of the efficacy of APD in the treatment of\ndelirium using retrospective cohort analysis. We plan to use the Causal\ninference framework to look for the underlying causal structure model,\nleveraging the availability of large observational data on ICU patients. To\nexplore safety outcomes associated with APD, we aim to build a causal model for\ndelirium in the ICU using large observational data sets connecting various\ncovariates correlated with delirium. We utilized the MIMIC III database, an\nextensive electronic health records (EHR) dataset with 53,423 distinct hospital\nadmissions. Our null hypothesis is: there is no significant difference in\noutcomes for delirium patients under different drug-group in the ICU. Through\nour exploratory, machine learning based and causal analysis, we had findings\nsuch as: mean length-of-stay and max length-of-stay is higher for patients in\nHaloperidol drug group, and haloperidol group has a higher rate of death in a\nyear compared to other two-groups. Our generated causal model explicitly shows\nthe functional relationships between different covariates. For future work, we\nplan to do time-varying analysis on the dataset.",
"title": "Causal Discovery on the Effect of Antipsychotic Drugs on Delirium Patients in the ICU using Large EHR Dataset",
"url": "http://arxiv.org/abs/2205.01057v1"
} | null | null | no_new_dataset | admin | null | false | null | 54e8fd8a-61e8-4ece-8a82-6aed4827a698 | null | Validated | 2023-10-04 15:19:51.886775 | {
"text_length": 1797
} | 1no_new_dataset
|
TITLE: XCOPA: A Multilingual Dataset for Causal Commonsense Reasoning
ABSTRACT: In order to simulate human language capacity, natural language processing
systems must be able to reason about the dynamics of everyday situations,
including their possible causes and effects. Moreover, they should be able to
generalise the acquired world knowledge to new languages, modulo cultural
differences. Advances in machine reasoning and cross-lingual transfer depend on
the availability of challenging evaluation benchmarks. Motivated by both
demands, we introduce Cross-lingual Choice of Plausible Alternatives (XCOPA), a
typologically diverse multilingual dataset for causal commonsense reasoning in
11 languages, which includes resource-poor languages like Eastern Apur\'imac
Quechua and Haitian Creole. We evaluate a range of state-of-the-art models on
this novel dataset, revealing that the performance of current methods based on
multilingual pretraining and zero-shot fine-tuning falls short compared to
translation-based transfer. Finally, we propose strategies to adapt
multilingual models to out-of-sample resource-lean languages where only a small
corpus or a bilingual dictionary is available, and report substantial
improvements over the random baseline. The XCOPA dataset is freely available at
github.com/cambridgeltl/xcopa. | {
"abstract": "In order to simulate human language capacity, natural language processing\nsystems must be able to reason about the dynamics of everyday situations,\nincluding their possible causes and effects. Moreover, they should be able to\ngeneralise the acquired world knowledge to new languages, modulo cultural\ndifferences. Advances in machine reasoning and cross-lingual transfer depend on\nthe availability of challenging evaluation benchmarks. Motivated by both\ndemands, we introduce Cross-lingual Choice of Plausible Alternatives (XCOPA), a\ntypologically diverse multilingual dataset for causal commonsense reasoning in\n11 languages, which includes resource-poor languages like Eastern Apur\\'imac\nQuechua and Haitian Creole. We evaluate a range of state-of-the-art models on\nthis novel dataset, revealing that the performance of current methods based on\nmultilingual pretraining and zero-shot fine-tuning falls short compared to\ntranslation-based transfer. Finally, we propose strategies to adapt\nmultilingual models to out-of-sample resource-lean languages where only a small\ncorpus or a bilingual dictionary is available, and report substantial\nimprovements over the random baseline. The XCOPA dataset is freely available at\ngithub.com/cambridgeltl/xcopa.",
"title": "XCOPA: A Multilingual Dataset for Causal Commonsense Reasoning",
"url": "http://arxiv.org/abs/2005.00333v2"
} | null | null | new_dataset | admin | null | false | null | 37687429-1667-4057-931e-c1fc4bf18ef2 | null | Validated | 2023-10-04 15:19:51.900328 | {
"text_length": 1346
} | 0new_dataset
|
TITLE: D4: a Chinese Dialogue Dataset for Depression-Diagnosis-Oriented Chat
ABSTRACT: In a depression-diagnosis-directed clinical session, doctors initiate a
conversation with ample emotional support that guides the patients to expose
their symptoms based on clinical diagnosis criteria. Such a dialogue system is
distinguished from existing single-purpose human-machine dialog systems, as it
combines task-oriented and chit-chats with uniqueness in dialogue topics and
procedures. However, due to the social stigma associated with mental illness,
the dialogue data related to depression consultation and diagnosis are rarely
disclosed. Based on clinical depression diagnostic criteria ICD-11 and DSM-5,
we designed a 3-phase procedure to construct D$^4$: a Chinese Dialogue Dataset
for Depression-Diagnosis-Oriented Chat, which simulates the dialogue between
doctors and patients during the diagnosis of depression, including diagnosis
results and symptom summary given by professional psychiatrists for each
conversation. Upon the newly-constructed dataset, four tasks mirroring the
depression diagnosis process are established: response generation, topic
prediction, dialog summary, and severity classification of depressive episode
and suicide risk. Multi-scale evaluation results demonstrate that a more
empathy-driven and diagnostic-accurate consultation dialogue system trained on
our dataset can be achieved compared to rule-based bots. | {
"abstract": "In a depression-diagnosis-directed clinical session, doctors initiate a\nconversation with ample emotional support that guides the patients to expose\ntheir symptoms based on clinical diagnosis criteria. Such a dialogue system is\ndistinguished from existing single-purpose human-machine dialog systems, as it\ncombines task-oriented and chit-chats with uniqueness in dialogue topics and\nprocedures. However, due to the social stigma associated with mental illness,\nthe dialogue data related to depression consultation and diagnosis are rarely\ndisclosed. Based on clinical depression diagnostic criteria ICD-11 and DSM-5,\nwe designed a 3-phase procedure to construct D$^4$: a Chinese Dialogue Dataset\nfor Depression-Diagnosis-Oriented Chat, which simulates the dialogue between\ndoctors and patients during the diagnosis of depression, including diagnosis\nresults and symptom summary given by professional psychiatrists for each\nconversation. Upon the newly-constructed dataset, four tasks mirroring the\ndepression diagnosis process are established: response generation, topic\nprediction, dialog summary, and severity classification of depressive episode\nand suicide risk. Multi-scale evaluation results demonstrate that a more\nempathy-driven and diagnostic-accurate consultation dialogue system trained on\nour dataset can be achieved compared to rule-based bots.",
"title": "D4: a Chinese Dialogue Dataset for Depression-Diagnosis-Oriented Chat",
"url": "http://arxiv.org/abs/2205.11764v2"
} | null | null | new_dataset | admin | null | false | null | db7cebff-6951-4c98-9923-77e2e937f9e3 | null | Validated | 2023-10-04 15:19:51.886231 | {
"text_length": 1462
} | 0new_dataset
|
TITLE: Application of quantum-inspired generative models to small molecular datasets
ABSTRACT: Quantum and quantum-inspired machine learning has emerged as a promising and
challenging research field due to the increased popularity of quantum
computing, especially with near-term devices. Theoretical contributions point
toward generative modeling as a promising direction to realize the first
examples of real-world quantum advantages from these technologies. A few
empirical studies also demonstrate such potential, especially when considering
quantum-inspired models based on tensor networks. In this work, we apply
tensor-network-based generative models to the problem of molecular discovery.
In our approach, we utilize two small molecular datasets: a subset of $4989$
molecules from the QM9 dataset and a small in-house dataset of $516$ validated
antioxidants from TotalEnergies. We compare several tensor network models
against a generative adversarial network using different sample-based metrics,
which reflect their learning performances on each task, and multiobjective
performances using $3$ relevant molecular metrics per task. We also combined
the output of the models and demonstrate empirically that such a combination
can be beneficial, advocating for the unification of classical and
quantum(-inspired) generative learning. | {
"abstract": "Quantum and quantum-inspired machine learning has emerged as a promising and\nchallenging research field due to the increased popularity of quantum\ncomputing, especially with near-term devices. Theoretical contributions point\ntoward generative modeling as a promising direction to realize the first\nexamples of real-world quantum advantages from these technologies. A few\nempirical studies also demonstrate such potential, especially when considering\nquantum-inspired models based on tensor networks. In this work, we apply\ntensor-network-based generative models to the problem of molecular discovery.\nIn our approach, we utilize two small molecular datasets: a subset of $4989$\nmolecules from the QM9 dataset and a small in-house dataset of $516$ validated\nantioxidants from TotalEnergies. We compare several tensor network models\nagainst a generative adversarial network using different sample-based metrics,\nwhich reflect their learning performances on each task, and multiobjective\nperformances using $3$ relevant molecular metrics per task. We also combined\nthe output of the models and demonstrate empirically that such a combination\ncan be beneficial, advocating for the unification of classical and\nquantum(-inspired) generative learning.",
"title": "Application of quantum-inspired generative models to small molecular datasets",
"url": "http://arxiv.org/abs/2304.10867v1"
} | null | null | no_new_dataset | admin | null | false | null | 3ea8c2bb-16ce-42ab-97fe-d385882130af | null | Validated | 2023-10-04 15:19:51.879611 | {
"text_length": 1357
} | 1no_new_dataset
|
TITLE: DatasetEquity: Are All Samples Created Equal? In The Quest For Equity Within Datasets
ABSTRACT: Data imbalance is a well-known issue in the field of machine learning,
attributable to the cost of data collection, the difficulty of labeling, and
the geographical distribution of the data. In computer vision, bias in data
distribution caused by image appearance remains highly unexplored. Compared to
categorical distributions using class labels, image appearance reveals complex
relationships between objects beyond what class labels provide. Clustering deep
perceptual features extracted from raw pixels gives a richer representation of
the data. This paper presents a novel method for addressing data imbalance in
machine learning. The method computes sample likelihoods based on image
appearance using deep perceptual embeddings and clustering. It then uses these
likelihoods to weigh samples differently during training with a proposed
$\textbf{Generalized Focal Loss}$ function. This loss can be easily integrated
with deep learning algorithms. Experiments validate the method's effectiveness
across autonomous driving vision datasets including KITTI and nuScenes. The
loss function improves state-of-the-art 3D object detection methods, achieving
over $200\%$ AP gains on under-represented classes (Cyclist) in the KITTI
dataset. The results demonstrate the method is generalizable, complements
existing techniques, and is particularly beneficial for smaller datasets and
rare classes. Code is available at:
https://github.com/towardsautonomy/DatasetEquity | {
"abstract": "Data imbalance is a well-known issue in the field of machine learning,\nattributable to the cost of data collection, the difficulty of labeling, and\nthe geographical distribution of the data. In computer vision, bias in data\ndistribution caused by image appearance remains highly unexplored. Compared to\ncategorical distributions using class labels, image appearance reveals complex\nrelationships between objects beyond what class labels provide. Clustering deep\nperceptual features extracted from raw pixels gives a richer representation of\nthe data. This paper presents a novel method for addressing data imbalance in\nmachine learning. The method computes sample likelihoods based on image\nappearance using deep perceptual embeddings and clustering. It then uses these\nlikelihoods to weigh samples differently during training with a proposed\n$\\textbf{Generalized Focal Loss}$ function. This loss can be easily integrated\nwith deep learning algorithms. Experiments validate the method's effectiveness\nacross autonomous driving vision datasets including KITTI and nuScenes. The\nloss function improves state-of-the-art 3D object detection methods, achieving\nover $200\\%$ AP gains on under-represented classes (Cyclist) in the KITTI\ndataset. The results demonstrate the method is generalizable, complements\nexisting techniques, and is particularly beneficial for smaller datasets and\nrare classes. Code is available at:\nhttps://github.com/towardsautonomy/DatasetEquity",
"title": "DatasetEquity: Are All Samples Created Equal? In The Quest For Equity Within Datasets",
"url": "http://arxiv.org/abs/2308.09878v2"
} | null | null | no_new_dataset | admin | null | false | null | 3770fe4a-da39-4be4-a2f3-f219bb05de59 | null | Validated | 2023-10-04 15:19:51.864060 | {
"text_length": 1585
} | 1no_new_dataset
|
TITLE: Introducing various Semantic Models for Amharic: Experimentation and Evaluation with multiple Tasks and Datasets
ABSTRACT: The availability of different pre-trained semantic models enabled the quick
development of machine learning components for downstream applications. Despite
the availability of abundant text data for low resource languages, only a few
semantic models are publicly available. Publicly available pre-trained models
are usually built as a multilingual version of semantic models that can not fit
well for each language due to context variations. In this work, we introduce
different semantic models for Amharic. After we experiment with the existing
pre-trained semantic models, we trained and fine-tuned nine new different
models using a monolingual text corpus. The models are build using word2Vec
embeddings, distributional thesaurus (DT), contextual embeddings, and DT
embeddings obtained via network embedding algorithms. Moreover, we employ these
models for different NLP tasks and investigate their impact. We find that newly
trained models perform better than pre-trained multilingual models.
Furthermore, models based on contextual embeddings from RoBERTA perform better
than the word2Vec models. | {
"abstract": "The availability of different pre-trained semantic models enabled the quick\ndevelopment of machine learning components for downstream applications. Despite\nthe availability of abundant text data for low resource languages, only a few\nsemantic models are publicly available. Publicly available pre-trained models\nare usually built as a multilingual version of semantic models that can not fit\nwell for each language due to context variations. In this work, we introduce\ndifferent semantic models for Amharic. After we experiment with the existing\npre-trained semantic models, we trained and fine-tuned nine new different\nmodels using a monolingual text corpus. The models are build using word2Vec\nembeddings, distributional thesaurus (DT), contextual embeddings, and DT\nembeddings obtained via network embedding algorithms. Moreover, we employ these\nmodels for different NLP tasks and investigate their impact. We find that newly\ntrained models perform better than pre-trained multilingual models.\nFurthermore, models based on contextual embeddings from RoBERTA perform better\nthan the word2Vec models.",
"title": "Introducing various Semantic Models for Amharic: Experimentation and Evaluation with multiple Tasks and Datasets",
"url": "http://arxiv.org/abs/2011.01154v2"
} | null | null | no_new_dataset | admin | null | false | null | 23f200e3-8943-4963-b563-044769105c27 | null | Validated | 2023-10-04 15:19:51.897283 | {
"text_length": 1248
} | 1no_new_dataset
|
TITLE: WeatherBench Probability: A benchmark dataset for probabilistic medium-range weather forecasting along with deep learning baseline models
ABSTRACT: WeatherBench is a benchmark dataset for medium-range weather forecasting of
geopotential, temperature and precipitation, consisting of preprocessed data,
predefined evaluation metrics and a number of baseline models. WeatherBench
Probability extends this to probabilistic forecasting by adding a set of
established probabilistic verification metrics (continuous ranked probability
score, spread-skill ratio and rank histograms) and a state-of-the-art
operational baseline using the ECWMF IFS ensemble forecast. In addition, we
test three different probabilistic machine learning methods -- Monte Carlo
dropout, parametric prediction and categorical prediction, in which the
probability distribution is discretized. We find that plain Monte Carlo dropout
severely underestimates uncertainty. The parametric and categorical models both
produce fairly reliable forecasts of similar quality. The parametric models
have fewer degrees of freedom while the categorical model is more flexible when
it comes to predicting non-Gaussian distributions. None of the models are able
to match the skill of the operational IFS model. We hope that this benchmark
will enable other researchers to evaluate their probabilistic approaches. | {
"abstract": "WeatherBench is a benchmark dataset for medium-range weather forecasting of\ngeopotential, temperature and precipitation, consisting of preprocessed data,\npredefined evaluation metrics and a number of baseline models. WeatherBench\nProbability extends this to probabilistic forecasting by adding a set of\nestablished probabilistic verification metrics (continuous ranked probability\nscore, spread-skill ratio and rank histograms) and a state-of-the-art\noperational baseline using the ECWMF IFS ensemble forecast. In addition, we\ntest three different probabilistic machine learning methods -- Monte Carlo\ndropout, parametric prediction and categorical prediction, in which the\nprobability distribution is discretized. We find that plain Monte Carlo dropout\nseverely underestimates uncertainty. The parametric and categorical models both\nproduce fairly reliable forecasts of similar quality. The parametric models\nhave fewer degrees of freedom while the categorical model is more flexible when\nit comes to predicting non-Gaussian distributions. None of the models are able\nto match the skill of the operational IFS model. We hope that this benchmark\nwill enable other researchers to evaluate their probabilistic approaches.",
"title": "WeatherBench Probability: A benchmark dataset for probabilistic medium-range weather forecasting along with deep learning baseline models",
"url": "http://arxiv.org/abs/2205.00865v1"
} | null | null | new_dataset | admin | null | false | null | 5bb9aec9-7d15-4ec3-b371-1c94c4e3fcbe | null | Validated | 2023-10-04 15:19:51.886704 | {
"text_length": 1391
} | 0new_dataset
|
TITLE: Systematic Atomic Structure Datasets for Machine Learning Potentials: Application to Defects in Magnesium
ABSTRACT: We present a physically motivated strategy for the construction of training
sets for transferable machine learning interatomic potentials. It is based on a
systematic exploration of all possible space groups in random crystal
structures, together with deformations of cell shape, size, and atomic
positions. The resulting potentials turn out to be unbiased and generically
applicable to studies of bulk defects without including any defect structures
in the training set or employing any additional Active Learning. Using this
approach we construct transferable potentials for pure Magnesium that reproduce
the properties of hexagonal closed packed (hcp) and body centered cubic (bcc)
polymorphs very well. In the process we investigate how different types of
training structures impact the properties and the predictive power of the
resulting potential. | {
"abstract": "We present a physically motivated strategy for the construction of training\nsets for transferable machine learning interatomic potentials. It is based on a\nsystematic exploration of all possible space groups in random crystal\nstructures, together with deformations of cell shape, size, and atomic\npositions. The resulting potentials turn out to be unbiased and generically\napplicable to studies of bulk defects without including any defect structures\nin the training set or employing any additional Active Learning. Using this\napproach we construct transferable potentials for pure Magnesium that reproduce\nthe properties of hexagonal closed packed (hcp) and body centered cubic (bcc)\npolymorphs very well. In the process we investigate how different types of\ntraining structures impact the properties and the predictive power of the\nresulting potential.",
"title": "Systematic Atomic Structure Datasets for Machine Learning Potentials: Application to Defects in Magnesium",
"url": "http://arxiv.org/abs/2207.04009v4"
} | null | null | no_new_dataset | admin | null | false | null | 1eedf0be-7129-4ee7-8207-6f14449ffc6f | null | Validated | 2023-10-04 15:19:51.885391 | {
"text_length": 994
} | 1no_new_dataset
|
TITLE: MD-HIT: Machine learning for materials property prediction with dataset redundancy control
ABSTRACT: Materials datasets are usually featured by the existence of many redundant
(highly similar) materials due to the tinkering material design practice over
the history of materials research. For example, the materials project database
has many perovskite cubic structure materials similar to SrTiO$_3$. This sample
redundancy within the dataset makes the random splitting of machine learning
model evaluation to fail so that the ML models tend to achieve over-estimated
predictive performance which is misleading for the materials science community.
This issue is well known in the field of bioinformatics for protein function
prediction, in which a redundancy reduction procedure (CD-Hit) is always
applied to reduce the sample redundancy by ensuring no pair of samples has a
sequence similarity greater than a given threshold. This paper surveys the
overestimated ML performance in the literature for both composition based and
structure based material property prediction. We then propose a material
dataset redundancy reduction algorithm called MD-HIT and evaluate it with
several composition and structure based distance threshold sfor reducing data
set sample redundancy. We show that with this control, the predicted
performance tends to better reflect their true prediction capability. Our
MD-hit code can be freely accessed at https://github.com/usccolumbia/MD-HIT | {
"abstract": "Materials datasets are usually featured by the existence of many redundant\n(highly similar) materials due to the tinkering material design practice over\nthe history of materials research. For example, the materials project database\nhas many perovskite cubic structure materials similar to SrTiO$_3$. This sample\nredundancy within the dataset makes the random splitting of machine learning\nmodel evaluation to fail so that the ML models tend to achieve over-estimated\npredictive performance which is misleading for the materials science community.\nThis issue is well known in the field of bioinformatics for protein function\nprediction, in which a redundancy reduction procedure (CD-Hit) is always\napplied to reduce the sample redundancy by ensuring no pair of samples has a\nsequence similarity greater than a given threshold. This paper surveys the\noverestimated ML performance in the literature for both composition based and\nstructure based material property prediction. We then propose a material\ndataset redundancy reduction algorithm called MD-HIT and evaluate it with\nseveral composition and structure based distance threshold sfor reducing data\nset sample redundancy. We show that with this control, the predicted\nperformance tends to better reflect their true prediction capability. Our\nMD-hit code can be freely accessed at https://github.com/usccolumbia/MD-HIT",
"title": "MD-HIT: Machine learning for materials property prediction with dataset redundancy control",
"url": "http://arxiv.org/abs/2307.04351v1"
} | null | null | no_new_dataset | admin | null | false | null | 5ef653b7-501b-4522-992c-7f3de80a5c62 | null | Validated | 2023-10-04 15:19:51.867758 | {
"text_length": 1495
} | 1no_new_dataset
|
TITLE: A Dataset for Evaluating Blood Detection in Hyperspectral Images
ABSTRACT: The sensitivity of imaging spectroscopy to haemoglobin derivatives makes it a
promising tool for detecting blood. However, due to complexity and high
dimensionality of hyperspectral images, the development of hyperspectral blood
detection algorithms is challenging. To facilitate their development, we
present a new hyperspectral blood detection dataset. This dataset, published in
accordance to open access mandate, consist of multiple detection scenarios with
varying levels of complexity. It allows to test the performance of Machine
Learning methods in relation to different acquisition environments, types of
background, age of blood and presence of other blood-like substances. We
explored the dataset with blood detection experiments. We used hyperspectral
target detection algorithm based on the well-known Matched Filter detector. Our
results and their discussion highlight the challenges of blood detection in
hyperspectral data and form a reference for further works. | {
"abstract": "The sensitivity of imaging spectroscopy to haemoglobin derivatives makes it a\npromising tool for detecting blood. However, due to complexity and high\ndimensionality of hyperspectral images, the development of hyperspectral blood\ndetection algorithms is challenging. To facilitate their development, we\npresent a new hyperspectral blood detection dataset. This dataset, published in\naccordance to open access mandate, consist of multiple detection scenarios with\nvarying levels of complexity. It allows to test the performance of Machine\nLearning methods in relation to different acquisition environments, types of\nbackground, age of blood and presence of other blood-like substances. We\nexplored the dataset with blood detection experiments. We used hyperspectral\ntarget detection algorithm based on the well-known Matched Filter detector. Our\nresults and their discussion highlight the challenges of blood detection in\nhyperspectral data and form a reference for further works.",
"title": "A Dataset for Evaluating Blood Detection in Hyperspectral Images",
"url": "http://arxiv.org/abs/2008.10254v2"
} | null | null | new_dataset | admin | null | false | null | 6550dd1c-4323-48ad-8000-b41218796947 | null | Validated | 2023-10-04 15:19:51.898584 | {
"text_length": 1077
} | 0new_dataset
|
TITLE: Deep image prior for 3D magnetic particle imaging: A quantitative comparison of regularization techniques on Open MPI dataset
ABSTRACT: Magnetic particle imaging (MPI) is an imaging modality exploiting the
nonlinear magnetization behavior of (super-)paramagnetic nanoparticles to
obtain a space- and often also time-dependent concentration of a tracer
consisting of these nanoparticles. MPI has a continuously increasing number of
potential medical applications. One prerequisite for successful performance in
these applications is a proper solution to the image reconstruction problem.
More classical methods from inverse problems theory, as well as novel
approaches from the field of machine learning, have the potential to deliver
high-quality reconstructions in MPI. We investigate a novel reconstruction
approach based on a deep image prior, which builds on representing the solution
by a deep neural network. Novel approaches, as well as variational and
iterative regularization techniques, are compared quantitatively in terms of
peak signal-to-noise ratios and structural similarity indices on the publicly
available Open MPI dataset. | {
"abstract": "Magnetic particle imaging (MPI) is an imaging modality exploiting the\nnonlinear magnetization behavior of (super-)paramagnetic nanoparticles to\nobtain a space- and often also time-dependent concentration of a tracer\nconsisting of these nanoparticles. MPI has a continuously increasing number of\npotential medical applications. One prerequisite for successful performance in\nthese applications is a proper solution to the image reconstruction problem.\nMore classical methods from inverse problems theory, as well as novel\napproaches from the field of machine learning, have the potential to deliver\nhigh-quality reconstructions in MPI. We investigate a novel reconstruction\napproach based on a deep image prior, which builds on representing the solution\nby a deep neural network. Novel approaches, as well as variational and\niterative regularization techniques, are compared quantitatively in terms of\npeak signal-to-noise ratios and structural similarity indices on the publicly\navailable Open MPI dataset.",
"title": "Deep image prior for 3D magnetic particle imaging: A quantitative comparison of regularization techniques on Open MPI dataset",
"url": "http://arxiv.org/abs/2007.01593v1"
} | null | null | no_new_dataset | admin | null | false | null | 65ae972e-b15b-46c8-a9b3-2684311c2738 | null | Validated | 2023-10-04 15:19:51.899267 | {
"text_length": 1166
} | 1no_new_dataset
|
TITLE: A Survey of COVID-19 Misinformation: Datasets, Detection Techniques and Open Issues
ABSTRACT: Misinformation during pandemic situations like COVID-19 is growing rapidly on
social media and other platforms. This expeditious growth of misinformation
creates adverse effects on the people living in the society. Researchers are
trying their best to mitigate this problem using different approaches based on
Machine Learning (ML), Deep Learning (DL), and Natural Language Processing
(NLP). This survey aims to study different approaches of misinformation
detection on COVID-19 in recent literature to help the researchers in this
domain. More specifically, we review the different methods used for COVID-19
misinformation detection in their research with an overview of data
pre-processing and feature extraction methods to get a better understanding of
their work. We also summarize the existing datasets which can be used for
further research. Finally, we discuss the limitations of the existing methods
and highlight some potential future research directions along this dimension to
combat the spreading of misinformation during a pandemic. | {
"abstract": "Misinformation during pandemic situations like COVID-19 is growing rapidly on\nsocial media and other platforms. This expeditious growth of misinformation\ncreates adverse effects on the people living in the society. Researchers are\ntrying their best to mitigate this problem using different approaches based on\nMachine Learning (ML), Deep Learning (DL), and Natural Language Processing\n(NLP). This survey aims to study different approaches of misinformation\ndetection on COVID-19 in recent literature to help the researchers in this\ndomain. More specifically, we review the different methods used for COVID-19\nmisinformation detection in their research with an overview of data\npre-processing and feature extraction methods to get a better understanding of\ntheir work. We also summarize the existing datasets which can be used for\nfurther research. Finally, we discuss the limitations of the existing methods\nand highlight some potential future research directions along this dimension to\ncombat the spreading of misinformation during a pandemic.",
"title": "A Survey of COVID-19 Misinformation: Datasets, Detection Techniques and Open Issues",
"url": "http://arxiv.org/abs/2110.00737v2"
} | null | null | no_new_dataset | admin | null | false | null | 47b46f44-7258-47c2-af0f-96e85601e505 | null | Validated | 2023-10-04 15:19:51.891754 | {
"text_length": 1163
} | 1no_new_dataset
|
TITLE: GreenDB -- A Dataset and Benchmark for Extraction of Sustainability Information of Consumer Goods
ABSTRACT: The production, shipping, usage, and disposal of consumer goods have a
substantial impact on greenhouse gas emissions and the depletion of resources.
Machine Learning (ML) can help to foster sustainable consumption patterns by
accounting for sustainability aspects in product search or recommendations of
modern retail platforms. However, the lack of large high quality publicly
available product data with trustworthy sustainability information impedes the
development of ML technology that can help to reach our sustainability goals.
Here we present GreenDB, a database that collects products from European online
shops on a weekly basis. As proxy for the products' sustainability, it relies
on sustainability labels, which are evaluated by experts. The GreenDB schema
extends the well-known schema.org Product definition and can be readily
integrated into existing product catalogs. We present initial results
demonstrating that ML models trained with our data can reliably (F1 score 96%)
predict the sustainability label of products. These contributions can help to
complement existing e-commerce experiences and ultimately encourage users to
more sustainable consumption patterns. | {
"abstract": "The production, shipping, usage, and disposal of consumer goods have a\nsubstantial impact on greenhouse gas emissions and the depletion of resources.\nMachine Learning (ML) can help to foster sustainable consumption patterns by\naccounting for sustainability aspects in product search or recommendations of\nmodern retail platforms. However, the lack of large high quality publicly\navailable product data with trustworthy sustainability information impedes the\ndevelopment of ML technology that can help to reach our sustainability goals.\nHere we present GreenDB, a database that collects products from European online\nshops on a weekly basis. As proxy for the products' sustainability, it relies\non sustainability labels, which are evaluated by experts. The GreenDB schema\nextends the well-known schema.org Product definition and can be readily\nintegrated into existing product catalogs. We present initial results\ndemonstrating that ML models trained with our data can reliably (F1 score 96%)\npredict the sustainability label of products. These contributions can help to\ncomplement existing e-commerce experiences and ultimately encourage users to\nmore sustainable consumption patterns.",
"title": "GreenDB -- A Dataset and Benchmark for Extraction of Sustainability Information of Consumer Goods",
"url": "http://arxiv.org/abs/2207.10733v3"
} | null | null | new_dataset | admin | null | false | null | ada2917c-cd1e-4e57-88c4-5fdc7c1de219 | null | Validated | 2023-10-04 15:19:51.885175 | {
"text_length": 1317
} | 0new_dataset
|
TITLE: CodeQA: A Question Answering Dataset for Source Code Comprehension
ABSTRACT: We propose CodeQA, a free-form question answering dataset for the purpose of
source code comprehension: given a code snippet and a question, a textual
answer is required to be generated. CodeQA contains a Java dataset with 119,778
question-answer pairs and a Python dataset with 70,085 question-answer pairs.
To obtain natural and faithful questions and answers, we implement syntactic
rules and semantic analysis to transform code comments into question-answer
pairs. We present the construction process and conduct systematic analysis of
our dataset. Experiment results achieved by several neural baselines on our
dataset are shown and discussed. While research on question-answering and
machine reading comprehension develops rapidly, few prior work has drawn
attention to code question answering. This new dataset can serve as a useful
research benchmark for source code comprehension. | {
"abstract": "We propose CodeQA, a free-form question answering dataset for the purpose of\nsource code comprehension: given a code snippet and a question, a textual\nanswer is required to be generated. CodeQA contains a Java dataset with 119,778\nquestion-answer pairs and a Python dataset with 70,085 question-answer pairs.\nTo obtain natural and faithful questions and answers, we implement syntactic\nrules and semantic analysis to transform code comments into question-answer\npairs. We present the construction process and conduct systematic analysis of\nour dataset. Experiment results achieved by several neural baselines on our\ndataset are shown and discussed. While research on question-answering and\nmachine reading comprehension develops rapidly, few prior work has drawn\nattention to code question answering. This new dataset can serve as a useful\nresearch benchmark for source code comprehension.",
"title": "CodeQA: A Question Answering Dataset for Source Code Comprehension",
"url": "http://arxiv.org/abs/2109.08365v1"
} | null | null | new_dataset | admin | null | false | null | 344dd92c-25c9-4a8e-981d-222e926465af | null | Validated | 2023-10-04 15:19:51.892092 | {
"text_length": 990
} | 0new_dataset
|
TITLE: A robust kernel machine regression towards biomarker selection in multi-omics datasets of osteoporosis for drug discovery
ABSTRACT: Many statistical machine approaches could ultimately highlight novel features
of the etiology of complex diseases by analyzing multi-omics data. However,
they are sensitive to some deviations in distribution when the observed samples
are potentially contaminated with adversarial corrupted outliers (e.g., a
fictional data distribution). Likewise, statistical advances lag in supporting
comprehensive data-driven analyses of complex multi-omics data integration. We
propose a novel non-linear M-estimator-based approach, "robust kernel machine
regression (RobKMR)," to improve the robustness of statistical machine
regression and the diversity of fictional data to examine the higher-order
composite effect of multi-omics datasets. We address a robust kernel-centered
Gram matrix to estimate the model parameters accurately. We also propose a
robust score test to assess the marginal and joint Hadamard product of features
from multi-omics data. We apply our proposed approach to a multi-omics dataset
of osteoporosis (OP) from Caucasian females. Experiments demonstrate that the
proposed approach effectively identifies the inter-related risk factors of OP.
With solid evidence (p-value = 0.00001), biological validations, network-based
analysis, causal inference, and drug repurposing, the selected three triplets
((DKK1, SMTN, DRGX), (MTND5, FASTKD2, CSMD3), (MTND5, COG3, CSMD3)) are
significant biomarkers and directly relate to BMD. Overall, the top three
selected genes (DKK1, MTND5, FASTKD2) and one gene (SIDT1 at p-value= 0.001)
significantly bond with four drugs- Tacrolimus, Ibandronate, Alendronate, and
Bazedoxifene out of 30 candidates for drug repurposing in OP. Further, the
proposed approach can be applied to any disease model where multi-omics
datasets are available. | {
"abstract": "Many statistical machine approaches could ultimately highlight novel features\nof the etiology of complex diseases by analyzing multi-omics data. However,\nthey are sensitive to some deviations in distribution when the observed samples\nare potentially contaminated with adversarial corrupted outliers (e.g., a\nfictional data distribution). Likewise, statistical advances lag in supporting\ncomprehensive data-driven analyses of complex multi-omics data integration. We\npropose a novel non-linear M-estimator-based approach, \"robust kernel machine\nregression (RobKMR),\" to improve the robustness of statistical machine\nregression and the diversity of fictional data to examine the higher-order\ncomposite effect of multi-omics datasets. We address a robust kernel-centered\nGram matrix to estimate the model parameters accurately. We also propose a\nrobust score test to assess the marginal and joint Hadamard product of features\nfrom multi-omics data. We apply our proposed approach to a multi-omics dataset\nof osteoporosis (OP) from Caucasian females. Experiments demonstrate that the\nproposed approach effectively identifies the inter-related risk factors of OP.\nWith solid evidence (p-value = 0.00001), biological validations, network-based\nanalysis, causal inference, and drug repurposing, the selected three triplets\n((DKK1, SMTN, DRGX), (MTND5, FASTKD2, CSMD3), (MTND5, COG3, CSMD3)) are\nsignificant biomarkers and directly relate to BMD. Overall, the top three\nselected genes (DKK1, MTND5, FASTKD2) and one gene (SIDT1 at p-value= 0.001)\nsignificantly bond with four drugs- Tacrolimus, Ibandronate, Alendronate, and\nBazedoxifene out of 30 candidates for drug repurposing in OP. Further, the\nproposed approach can be applied to any disease model where multi-omics\ndatasets are available.",
"title": "A robust kernel machine regression towards biomarker selection in multi-omics datasets of osteoporosis for drug discovery",
"url": "http://arxiv.org/abs/2201.05060v1"
} | null | null | no_new_dataset | admin | null | false | null | 501d4478-7a1e-44ad-8efb-812daa2e7b8d | null | Validated | 2023-10-04 15:19:51.888867 | {
"text_length": 1943
} | 1no_new_dataset
|
TITLE: FLamby: Datasets and Benchmarks for Cross-Silo Federated Learning in Realistic Healthcare Settings
ABSTRACT: Federated Learning (FL) is a novel approach enabling several clients holding
sensitive data to collaboratively train machine learning models, without
centralizing data. The cross-silo FL setting corresponds to the case of few
($2$--$50$) reliable clients, each holding medium to large datasets, and is
typically found in applications such as healthcare, finance, or industry. While
previous works have proposed representative datasets for cross-device FL, few
realistic healthcare cross-silo FL datasets exist, thereby slowing algorithmic
research in this critical application. In this work, we propose a novel
cross-silo dataset suite focused on healthcare, FLamby (Federated Learning
AMple Benchmark of Your cross-silo strategies), to bridge the gap between
theory and practice of cross-silo FL. FLamby encompasses 7 healthcare datasets
with natural splits, covering multiple tasks, modalities, and data volumes,
each accompanied with baseline training code. As an illustration, we
additionally benchmark standard FL algorithms on all datasets. Our flexible and
modular suite allows researchers to easily download datasets, reproduce results
and re-use the different components for their research. FLamby is available
at~\url{www.github.com/owkin/flamby}. | {
"abstract": "Federated Learning (FL) is a novel approach enabling several clients holding\nsensitive data to collaboratively train machine learning models, without\ncentralizing data. The cross-silo FL setting corresponds to the case of few\n($2$--$50$) reliable clients, each holding medium to large datasets, and is\ntypically found in applications such as healthcare, finance, or industry. While\nprevious works have proposed representative datasets for cross-device FL, few\nrealistic healthcare cross-silo FL datasets exist, thereby slowing algorithmic\nresearch in this critical application. In this work, we propose a novel\ncross-silo dataset suite focused on healthcare, FLamby (Federated Learning\nAMple Benchmark of Your cross-silo strategies), to bridge the gap between\ntheory and practice of cross-silo FL. FLamby encompasses 7 healthcare datasets\nwith natural splits, covering multiple tasks, modalities, and data volumes,\neach accompanied with baseline training code. As an illustration, we\nadditionally benchmark standard FL algorithms on all datasets. Our flexible and\nmodular suite allows researchers to easily download datasets, reproduce results\nand re-use the different components for their research. FLamby is available\nat~\\url{www.github.com/owkin/flamby}.",
"title": "FLamby: Datasets and Benchmarks for Cross-Silo Federated Learning in Realistic Healthcare Settings",
"url": "http://arxiv.org/abs/2210.04620v3"
} | null | null | new_dataset | admin | null | false | null | ef435b18-6113-4a62-9193-f703169cae46 | null | Validated | 2023-10-04 15:19:51.883616 | {
"text_length": 1390
} | 0new_dataset
|
TITLE: "Excavating AI" Re-excavated: Debunking a Fallacious Account of the JAFFE Dataset
ABSTRACT: Twenty-five years ago, my colleagues Miyuki Kamachi and Jiro Gyoba and I
designed and photographed JAFFE, a set of facial expression images intended for
use in a study of face perception. In 2019, without seeking permission or
informing us, Kate Crawford and Trevor Paglen exhibited JAFFE in two widely
publicized art shows. In addition, they published a nonfactual account of the
images in the essay "Excavating AI: The Politics of Images in Machine Learning
Training Sets." The present article recounts the creation of the JAFFE dataset
and unravels each of Crawford and Paglen's fallacious statements. I also
discuss JAFFE more broadly in connection with research on facial expression,
affective computing, and human-computer interaction. | {
"abstract": "Twenty-five years ago, my colleagues Miyuki Kamachi and Jiro Gyoba and I\ndesigned and photographed JAFFE, a set of facial expression images intended for\nuse in a study of face perception. In 2019, without seeking permission or\ninforming us, Kate Crawford and Trevor Paglen exhibited JAFFE in two widely\npublicized art shows. In addition, they published a nonfactual account of the\nimages in the essay \"Excavating AI: The Politics of Images in Machine Learning\nTraining Sets.\" The present article recounts the creation of the JAFFE dataset\nand unravels each of Crawford and Paglen's fallacious statements. I also\ndiscuss JAFFE more broadly in connection with research on facial expression,\naffective computing, and human-computer interaction.",
"title": "\"Excavating AI\" Re-excavated: Debunking a Fallacious Account of the JAFFE Dataset",
"url": "http://arxiv.org/abs/2107.13998v1"
} | null | null | no_new_dataset | admin | null | false | null | 65a54dfd-e864-4411-a93a-1083abd7413d | null | Validated | 2023-10-04 15:19:51.893430 | {
"text_length": 857
} | 1no_new_dataset
|
TITLE: MapInWild: A Remote Sensing Dataset to Address the Question What Makes Nature Wild
ABSTRACT: Antrophonegic pressure (i.e. human influence) on the environment is one of
the largest causes of the loss of biological diversity. Wilderness areas, in
contrast, are home to undisturbed ecological processes. However, there is no
biophysical definition of the term wilderness. Instead, wilderness is more of a
philosophical or cultural concept and thus cannot be easily delineated or
categorized in a technical manner. With this paper, (i) we introduce the task
of wilderness mapping by means of machine learning applied to satellite imagery
(ii) and publish MapInWild, a large-scale benchmark dataset curated for that
task. MapInWild is a multi-modal dataset and comprises various geodata acquired
and formed from a diverse set of Earth observation sensors. The dataset
consists of 8144 images with a shape of 1920 x 1920 pixels and is approximately
350 GB in size. The images are weakly annotated with three classes derived from
the World Database of Protected Areas - Strict Nature Reserves, Wilderness
Areas, and National Parks. With the dataset, which shall serve as a testbed for
developments in fields such as explainable machine learning and environmental
remote sensing, we hope to contribute to a deepening of our understanding of
the question "What makes nature wild?". | {
"abstract": "Antrophonegic pressure (i.e. human influence) on the environment is one of\nthe largest causes of the loss of biological diversity. Wilderness areas, in\ncontrast, are home to undisturbed ecological processes. However, there is no\nbiophysical definition of the term wilderness. Instead, wilderness is more of a\nphilosophical or cultural concept and thus cannot be easily delineated or\ncategorized in a technical manner. With this paper, (i) we introduce the task\nof wilderness mapping by means of machine learning applied to satellite imagery\n(ii) and publish MapInWild, a large-scale benchmark dataset curated for that\ntask. MapInWild is a multi-modal dataset and comprises various geodata acquired\nand formed from a diverse set of Earth observation sensors. The dataset\nconsists of 8144 images with a shape of 1920 x 1920 pixels and is approximately\n350 GB in size. The images are weakly annotated with three classes derived from\nthe World Database of Protected Areas - Strict Nature Reserves, Wilderness\nAreas, and National Parks. With the dataset, which shall serve as a testbed for\ndevelopments in fields such as explainable machine learning and environmental\nremote sensing, we hope to contribute to a deepening of our understanding of\nthe question \"What makes nature wild?\".",
"title": "MapInWild: A Remote Sensing Dataset to Address the Question What Makes Nature Wild",
"url": "http://arxiv.org/abs/2212.02265v1"
} | null | null | new_dataset | admin | null | false | null | 6ea5765d-6ead-4e79-b06f-dc0ffcfa9fc1 | null | Validated | 2023-10-04 15:19:51.882392 | {
"text_length": 1396
} | 0new_dataset
|
TITLE: DC-BENCH: Dataset Condensation Benchmark
ABSTRACT: Dataset Condensation is a newly emerging technique aiming at learning a tiny
dataset that captures the rich information encoded in the original dataset. As
the size of datasets contemporary machine learning models rely on becomes
increasingly large, condensation methods become a prominent direction for
accelerating network training and reducing data storage. Despite numerous
methods have been proposed in this rapidly growing field, evaluating and
comparing different condensation methods is non-trivial and still remains an
open issue. The quality of condensed dataset are often shadowed by many
critical contributing factors to the end performance, such as data augmentation
and model architectures. The lack of a systematic way to evaluate and compare
condensation methods not only hinders our understanding of existing techniques,
but also discourages practical usage of the synthesized datasets. This work
provides the first large-scale standardized benchmark on Dataset Condensation.
It consists of a suite of evaluations to comprehensively reflect the
generability and effectiveness of condensation methods through the lens of
their generated dataset. Leveraging this benchmark, we conduct a large-scale
study of current condensation methods, and report many insightful findings that
open up new possibilities for future development. The benchmark library,
including evaluators, baseline methods, and generated datasets, is open-sourced
to facilitate future research and application. | {
"abstract": "Dataset Condensation is a newly emerging technique aiming at learning a tiny\ndataset that captures the rich information encoded in the original dataset. As\nthe size of datasets contemporary machine learning models rely on becomes\nincreasingly large, condensation methods become a prominent direction for\naccelerating network training and reducing data storage. Despite numerous\nmethods have been proposed in this rapidly growing field, evaluating and\ncomparing different condensation methods is non-trivial and still remains an\nopen issue. The quality of condensed dataset are often shadowed by many\ncritical contributing factors to the end performance, such as data augmentation\nand model architectures. The lack of a systematic way to evaluate and compare\ncondensation methods not only hinders our understanding of existing techniques,\nbut also discourages practical usage of the synthesized datasets. This work\nprovides the first large-scale standardized benchmark on Dataset Condensation.\nIt consists of a suite of evaluations to comprehensively reflect the\ngenerability and effectiveness of condensation methods through the lens of\ntheir generated dataset. Leveraging this benchmark, we conduct a large-scale\nstudy of current condensation methods, and report many insightful findings that\nopen up new possibilities for future development. The benchmark library,\nincluding evaluators, baseline methods, and generated datasets, is open-sourced\nto facilitate future research and application.",
"title": "DC-BENCH: Dataset Condensation Benchmark",
"url": "http://arxiv.org/abs/2207.09639v2"
} | null | null | new_dataset | admin | null | false | null | 7e989614-89c1-4882-857d-88aa6cdbb468 | null | Validated | 2023-10-04 15:19:51.885247 | {
"text_length": 1568
} | 0new_dataset
|
TITLE: Simplistic Collection and Labeling Practices Limit the Utility of Benchmark Datasets for Twitter Bot Detection
ABSTRACT: Accurate bot detection is necessary for the safety and integrity of online
platforms. It is also crucial for research on the influence of bots in
elections, the spread of misinformation, and financial market manipulation.
Platforms deploy infrastructure to flag or remove automated accounts, but their
tools and data are not publicly available. Thus, the public must rely on
third-party bot detection. These tools employ machine learning and often
achieve near perfect performance for classification on existing datasets,
suggesting bot detection is accurate, reliable and fit for use in downstream
applications. We provide evidence that this is not the case and show that high
performance is attributable to limitations in dataset collection and labeling
rather than sophistication of the tools. Specifically, we show that simple
decision rules -- shallow decision trees trained on a small number of features
-- achieve near-state-of-the-art performance on most available datasets and
that bot detection datasets, even when combined together, do not generalize
well to out-of-sample datasets. Our findings reveal that predictions are highly
dependent on each dataset's collection and labeling procedures rather than
fundamental differences between bots and humans. These results have important
implications for both transparency in sampling and labeling procedures and
potential biases in research using existing bot detection tools for
pre-processing. | {
"abstract": "Accurate bot detection is necessary for the safety and integrity of online\nplatforms. It is also crucial for research on the influence of bots in\nelections, the spread of misinformation, and financial market manipulation.\nPlatforms deploy infrastructure to flag or remove automated accounts, but their\ntools and data are not publicly available. Thus, the public must rely on\nthird-party bot detection. These tools employ machine learning and often\nachieve near perfect performance for classification on existing datasets,\nsuggesting bot detection is accurate, reliable and fit for use in downstream\napplications. We provide evidence that this is not the case and show that high\nperformance is attributable to limitations in dataset collection and labeling\nrather than sophistication of the tools. Specifically, we show that simple\ndecision rules -- shallow decision trees trained on a small number of features\n-- achieve near-state-of-the-art performance on most available datasets and\nthat bot detection datasets, even when combined together, do not generalize\nwell to out-of-sample datasets. Our findings reveal that predictions are highly\ndependent on each dataset's collection and labeling procedures rather than\nfundamental differences between bots and humans. These results have important\nimplications for both transparency in sampling and labeling procedures and\npotential biases in research using existing bot detection tools for\npre-processing.",
"title": "Simplistic Collection and Labeling Practices Limit the Utility of Benchmark Datasets for Twitter Bot Detection",
"url": "http://arxiv.org/abs/2301.07015v2"
} | null | null | no_new_dataset | admin | null | false | null | 85b4854c-a3e4-43ab-ab1a-1cc9b46e8d43 | null | Validated | 2023-10-04 15:19:51.881461 | {
"text_length": 1598
} | 1no_new_dataset
|
TITLE: Unleashing Realistic Air Quality Forecasting: Introducing the Ready-to-Use PurpleAirSF Dataset
ABSTRACT: Air quality forecasting has garnered significant attention recently, with
data-driven models taking center stage due to advancements in machine learning
and deep learning models. However, researchers face challenges with complex
data acquisition and the lack of open-sourced datasets, hindering efficient
model validation. This paper introduces PurpleAirSF, a comprehensive and easily
accessible dataset collected from the PurpleAir network. With its high temporal
resolution, various air quality measures, and diverse geographical coverage,
this dataset serves as a useful tool for researchers aiming to develop novel
forecasting models, study air pollution patterns, and investigate their impacts
on health and the environment. We present a detailed account of the data
collection and processing methods employed to build PurpleAirSF. Furthermore,
we conduct preliminary experiments using both classic and modern
spatio-temporal forecasting models, thereby establishing a benchmark for future
air quality forecasting tasks. | {
"abstract": "Air quality forecasting has garnered significant attention recently, with\ndata-driven models taking center stage due to advancements in machine learning\nand deep learning models. However, researchers face challenges with complex\ndata acquisition and the lack of open-sourced datasets, hindering efficient\nmodel validation. This paper introduces PurpleAirSF, a comprehensive and easily\naccessible dataset collected from the PurpleAir network. With its high temporal\nresolution, various air quality measures, and diverse geographical coverage,\nthis dataset serves as a useful tool for researchers aiming to develop novel\nforecasting models, study air pollution patterns, and investigate their impacts\non health and the environment. We present a detailed account of the data\ncollection and processing methods employed to build PurpleAirSF. Furthermore,\nwe conduct preliminary experiments using both classic and modern\nspatio-temporal forecasting models, thereby establishing a benchmark for future\nair quality forecasting tasks.",
"title": "Unleashing Realistic Air Quality Forecasting: Introducing the Ready-to-Use PurpleAirSF Dataset",
"url": "http://arxiv.org/abs/2306.13948v1"
} | null | null | no_new_dataset | admin | null | false | null | 981780f0-a6b3-4f1c-9b2f-156e495c4cf3 | null | Validated | 2023-10-04 15:19:51.870062 | {
"text_length": 1154
} | 1no_new_dataset
|
TITLE: DISC: a Dataset for Integrated Sensing and Communication in mmWave Systems
ABSTRACT: In this paper we present DISC, a dataset of millimeter-wave channel impulse
response measurements for integrated human activity sensing and communication.
This is the first dataset collected with a software-defined radio testbed that
transmits 60 GHz IEEE 802-11ay-compliant packets and estimates the channel
response including reflections of the signal on the moving body parts of
subjects moving in an indoor environment. The provided data contain the
contribution of 7 subjects performing 4 different activities. Differently from
available radar-based millimeter-wave sensing datasets, our measurements are
collected using both uniform packet transmission times and sparse traffic
patterns from real Wi-Fi deployments. Thanks to these unique characteristics,
DISC serves as a multi-purpose benchmarking tool for machine learning-based
human activity recognition, radio frequency gait analysis, and sparse sensing
algorithms for next-generation integrated sensing and communication. | {
"abstract": "In this paper we present DISC, a dataset of millimeter-wave channel impulse\nresponse measurements for integrated human activity sensing and communication.\nThis is the first dataset collected with a software-defined radio testbed that\ntransmits 60 GHz IEEE 802-11ay-compliant packets and estimates the channel\nresponse including reflections of the signal on the moving body parts of\nsubjects moving in an indoor environment. The provided data contain the\ncontribution of 7 subjects performing 4 different activities. Differently from\navailable radar-based millimeter-wave sensing datasets, our measurements are\ncollected using both uniform packet transmission times and sparse traffic\npatterns from real Wi-Fi deployments. Thanks to these unique characteristics,\nDISC serves as a multi-purpose benchmarking tool for machine learning-based\nhuman activity recognition, radio frequency gait analysis, and sparse sensing\nalgorithms for next-generation integrated sensing and communication.",
"title": "DISC: a Dataset for Integrated Sensing and Communication in mmWave Systems",
"url": "http://arxiv.org/abs/2306.09469v1"
} | null | null | new_dataset | admin | null | false | null | 63b45844-1865-445b-8ac6-5cda54144d14 | null | Validated | 2023-10-04 15:19:51.870697 | {
"text_length": 1093
} | 0new_dataset
|
TITLE: A Dataset and Benchmark for Malaria Life-Cycle Classification in Thin Blood Smear Images
ABSTRACT: Malaria microscopy, microscopic examination of stained blood slides to detect
parasite Plasmodium, is considered to be a gold-standard for detecting
life-threatening disease malaria. Detecting the plasmodium parasite requires a
skilled examiner and may take up to 10 to 15 minutes to completely go through
the whole slide. Due to a lack of skilled medical professionals in the
underdeveloped or resource deficient regions, many cases go misdiagnosed;
resulting in unavoidable complications and/or undue medication. We propose to
complement the medical professionals by creating a deep learning-based method
to automatically detect (localize) the plasmodium parasites in the photograph
of stained film. To handle the unbalanced nature of the dataset, we adopt a
two-stage approach. Where the first stage is trained to detect blood cells and
classify them into just healthy or infected. The second stage is trained to
classify each detected cell further into the life-cycle stage. To facilitate
the research in machine learning-based malaria microscopy, we introduce a new
large scale microscopic image malaria dataset. Thirty-eight thousand cells are
tagged from the 345 microscopic images of different Giemsa-stained slides of
blood samples. Extensive experimentation is performed using different CNN
backbones including VGG, DenseNet, and ResNet on this dataset. Our experiments
and analysis reveal that the two-stage approach works better than the one-stage
approach for malaria detection. To ensure the usability of our approach, we
have also developed a mobile app that will be used by local hospitals for
investigation and educational purposes. The dataset, its annotations, and
implementation codes will be released upon publication of the paper. | {
"abstract": "Malaria microscopy, microscopic examination of stained blood slides to detect\nparasite Plasmodium, is considered to be a gold-standard for detecting\nlife-threatening disease malaria. Detecting the plasmodium parasite requires a\nskilled examiner and may take up to 10 to 15 minutes to completely go through\nthe whole slide. Due to a lack of skilled medical professionals in the\nunderdeveloped or resource deficient regions, many cases go misdiagnosed;\nresulting in unavoidable complications and/or undue medication. We propose to\ncomplement the medical professionals by creating a deep learning-based method\nto automatically detect (localize) the plasmodium parasites in the photograph\nof stained film. To handle the unbalanced nature of the dataset, we adopt a\ntwo-stage approach. Where the first stage is trained to detect blood cells and\nclassify them into just healthy or infected. The second stage is trained to\nclassify each detected cell further into the life-cycle stage. To facilitate\nthe research in machine learning-based malaria microscopy, we introduce a new\nlarge scale microscopic image malaria dataset. Thirty-eight thousand cells are\ntagged from the 345 microscopic images of different Giemsa-stained slides of\nblood samples. Extensive experimentation is performed using different CNN\nbackbones including VGG, DenseNet, and ResNet on this dataset. Our experiments\nand analysis reveal that the two-stage approach works better than the one-stage\napproach for malaria detection. To ensure the usability of our approach, we\nhave also developed a mobile app that will be used by local hospitals for\ninvestigation and educational purposes. The dataset, its annotations, and\nimplementation codes will be released upon publication of the paper.",
"title": "A Dataset and Benchmark for Malaria Life-Cycle Classification in Thin Blood Smear Images",
"url": "http://arxiv.org/abs/2102.08708v1"
} | null | null | new_dataset | admin | null | false | null | 3d753e99-34f4-4ee3-bcb4-c684c57dacb5 | null | Validated | 2023-10-04 15:19:51.895732 | {
"text_length": 1875
} | 0new_dataset
|
TITLE: TpuGraphs: A Performance Prediction Dataset on Large Tensor Computational Graphs
ABSTRACT: Precise hardware performance models play a crucial role in code
optimizations. They can assist compilers in making heuristic decisions or aid
autotuners in identifying the optimal configuration for a given program. For
example, the autotuner for XLA, a machine learning compiler, discovered 10-20%
speedup on state-of-the-art models serving substantial production traffic at
Google. Although there exist a few datasets for program performance prediction,
they target small sub-programs such as basic blocks or kernels. This paper
introduces TpuGraphs, a performance prediction dataset on full tensor programs,
represented as computational graphs, running on Tensor Processing Units (TPUs).
Each graph in the dataset represents the main computation of a machine learning
workload, e.g., a training epoch or an inference step. Each data sample
contains a computational graph, a compilation configuration, and the execution
time of the graph when compiled with the configuration. The graphs in the
dataset are collected from open-source machine learning programs, featuring
popular model architectures, e.g., ResNet, EfficientNet, Mask R-CNN, and
Transformer. TpuGraphs provides 25x more graphs than the largest graph property
prediction dataset (with comparable graph sizes), and 770x larger graphs on
average compared to existing performance prediction datasets on machine
learning programs. This graph-level prediction task on large graphs introduces
new challenges in learning, ranging from scalability, training efficiency, to
model quality. | {
"abstract": "Precise hardware performance models play a crucial role in code\noptimizations. They can assist compilers in making heuristic decisions or aid\nautotuners in identifying the optimal configuration for a given program. For\nexample, the autotuner for XLA, a machine learning compiler, discovered 10-20%\nspeedup on state-of-the-art models serving substantial production traffic at\nGoogle. Although there exist a few datasets for program performance prediction,\nthey target small sub-programs such as basic blocks or kernels. This paper\nintroduces TpuGraphs, a performance prediction dataset on full tensor programs,\nrepresented as computational graphs, running on Tensor Processing Units (TPUs).\nEach graph in the dataset represents the main computation of a machine learning\nworkload, e.g., a training epoch or an inference step. Each data sample\ncontains a computational graph, a compilation configuration, and the execution\ntime of the graph when compiled with the configuration. The graphs in the\ndataset are collected from open-source machine learning programs, featuring\npopular model architectures, e.g., ResNet, EfficientNet, Mask R-CNN, and\nTransformer. TpuGraphs provides 25x more graphs than the largest graph property\nprediction dataset (with comparable graph sizes), and 770x larger graphs on\naverage compared to existing performance prediction datasets on machine\nlearning programs. This graph-level prediction task on large graphs introduces\nnew challenges in learning, ranging from scalability, training efficiency, to\nmodel quality.",
"title": "TpuGraphs: A Performance Prediction Dataset on Large Tensor Computational Graphs",
"url": "http://arxiv.org/abs/2308.13490v1"
} | null | null | new_dataset | admin | null | false | null | 63548cbf-df52-4b72-9fc4-ed34470efa80 | null | Validated | 2023-10-04 15:19:51.863985 | {
"text_length": 1658
} | 0new_dataset
|
TITLE: The Open Catalyst 2022 (OC22) Dataset and Challenges for Oxide Electrocatalysts
ABSTRACT: The development of machine learning models for electrocatalysts requires a
broad set of training data to enable their use across a wide variety of
materials. One class of materials that currently lacks sufficient training data
is oxides, which are critical for the development of OER catalysts. To address
this, we developed the OC22 dataset, consisting of 62,331 DFT relaxations
(~9,854,504 single point calculations) across a range of oxide materials,
coverages, and adsorbates. We define generalized total energy tasks that enable
property prediction beyond adsorption energies; we test baseline performance of
several graph neural networks; and we provide pre-defined dataset splits to
establish clear benchmarks for future efforts. In the most general task,
GemNet-OC sees a ~36% improvement in energy predictions when combining the
chemically dissimilar OC20 and OC22 datasets via fine-tuning. Similarly, we
achieved a ~19% improvement in total energy predictions on OC20 and a ~9%
improvement in force predictions in OC22 when using joint training. We
demonstrate the practical utility of a top performing model by capturing
literature adsorption energies and important OER scaling relationships. We
expect OC22 to provide an important benchmark for models seeking to incorporate
intricate long-range electrostatic and magnetic interactions in oxide surfaces.
Dataset and baseline models are open sourced, and a public leaderboard is
available to encourage continued community developments on the total energy
tasks and data. | {
"abstract": "The development of machine learning models for electrocatalysts requires a\nbroad set of training data to enable their use across a wide variety of\nmaterials. One class of materials that currently lacks sufficient training data\nis oxides, which are critical for the development of OER catalysts. To address\nthis, we developed the OC22 dataset, consisting of 62,331 DFT relaxations\n(~9,854,504 single point calculations) across a range of oxide materials,\ncoverages, and adsorbates. We define generalized total energy tasks that enable\nproperty prediction beyond adsorption energies; we test baseline performance of\nseveral graph neural networks; and we provide pre-defined dataset splits to\nestablish clear benchmarks for future efforts. In the most general task,\nGemNet-OC sees a ~36% improvement in energy predictions when combining the\nchemically dissimilar OC20 and OC22 datasets via fine-tuning. Similarly, we\nachieved a ~19% improvement in total energy predictions on OC20 and a ~9%\nimprovement in force predictions in OC22 when using joint training. We\ndemonstrate the practical utility of a top performing model by capturing\nliterature adsorption energies and important OER scaling relationships. We\nexpect OC22 to provide an important benchmark for models seeking to incorporate\nintricate long-range electrostatic and magnetic interactions in oxide surfaces.\nDataset and baseline models are open sourced, and a public leaderboard is\navailable to encourage continued community developments on the total energy\ntasks and data.",
"title": "The Open Catalyst 2022 (OC22) Dataset and Challenges for Oxide Electrocatalysts",
"url": "http://arxiv.org/abs/2206.08917v3"
} | null | null | new_dataset | admin | null | false | null | 165e43bb-eb0b-4bc5-a64e-a4cdf40f12c2 | null | Validated | 2023-10-04 15:19:51.885677 | {
"text_length": 1646
} | 0new_dataset
|
TITLE: ToxiGen: A Large-Scale Machine-Generated Dataset for Adversarial and Implicit Hate Speech Detection
ABSTRACT: Toxic language detection systems often falsely flag text that contains
minority group mentions as toxic, as those groups are often the targets of
online hate. Such over-reliance on spurious correlations also causes systems to
struggle with detecting implicitly toxic language. To help mitigate these
issues, we create ToxiGen, a new large-scale and machine-generated dataset of
274k toxic and benign statements about 13 minority groups. We develop a
demonstration-based prompting framework and an adversarial
classifier-in-the-loop decoding method to generate subtly toxic and benign text
with a massive pretrained language model. Controlling machine generation in
this way allows ToxiGen to cover implicitly toxic text at a larger scale, and
about more demographic groups, than previous resources of human-written text.
We conduct a human evaluation on a challenging subset of ToxiGen and find that
annotators struggle to distinguish machine-generated text from human-written
language. We also find that 94.5% of toxic examples are labeled as hate speech
by human annotators. Using three publicly-available datasets, we show that
finetuning a toxicity classifier on our data improves its performance on
human-written data substantially. We also demonstrate that ToxiGen can be used
to fight machine-generated toxicity as finetuning improves the classifier
significantly on our evaluation subset. Our code and data can be found at
https://github.com/microsoft/ToxiGen. | {
"abstract": "Toxic language detection systems often falsely flag text that contains\nminority group mentions as toxic, as those groups are often the targets of\nonline hate. Such over-reliance on spurious correlations also causes systems to\nstruggle with detecting implicitly toxic language. To help mitigate these\nissues, we create ToxiGen, a new large-scale and machine-generated dataset of\n274k toxic and benign statements about 13 minority groups. We develop a\ndemonstration-based prompting framework and an adversarial\nclassifier-in-the-loop decoding method to generate subtly toxic and benign text\nwith a massive pretrained language model. Controlling machine generation in\nthis way allows ToxiGen to cover implicitly toxic text at a larger scale, and\nabout more demographic groups, than previous resources of human-written text.\nWe conduct a human evaluation on a challenging subset of ToxiGen and find that\nannotators struggle to distinguish machine-generated text from human-written\nlanguage. We also find that 94.5% of toxic examples are labeled as hate speech\nby human annotators. Using three publicly-available datasets, we show that\nfinetuning a toxicity classifier on our data improves its performance on\nhuman-written data substantially. We also demonstrate that ToxiGen can be used\nto fight machine-generated toxicity as finetuning improves the classifier\nsignificantly on our evaluation subset. Our code and data can be found at\nhttps://github.com/microsoft/ToxiGen.",
"title": "ToxiGen: A Large-Scale Machine-Generated Dataset for Adversarial and Implicit Hate Speech Detection",
"url": "http://arxiv.org/abs/2203.09509v4"
} | null | null | new_dataset | admin | null | false | null | 1e5d27d3-6715-439c-91d7-84723a31bca6 | null | Validated | 2023-10-04 15:19:51.887593 | {
"text_length": 1602
} | 0new_dataset
|
TITLE: A Wireless-Vision Dataset for Privacy Preserving Human Activity Recognition
ABSTRACT: Human Activity Recognition (HAR) has recently received remarkable attention
in numerous applications such as assisted living and remote monitoring.
Existing solutions based on sensors and vision technologies have obtained
achievements but still suffering from considerable limitations in the
environmental requirement. Wireless signals like WiFi-based sensing have
emerged as a new paradigm since it is convenient and not restricted in the
environment. In this paper, a new WiFi-based and video-based neural network
(WiNN) is proposed to improve the robustness of activity recognition where the
synchronized video serves as the supplement for the wireless data. Moreover, a
wireless-vision benchmark (WiVi) is collected for 9 class actions recognition
in three different visual conditions, including the scenes without occlusion,
with partial occlusion, and with full occlusion. Both machine learning methods
- support vector machine (SVM) as well as deep learning methods are used for
the accuracy verification of the data set. Our results show that WiVi data set
satisfies the primary demand and all three branches in the proposed pipeline
keep more than $80\%$ of activity recognition accuracy over multiple action
segmentation from 1s to 3s. In particular, WiNN is the most robust method in
terms of all the actions on three action segmentation compared to the others. | {
"abstract": "Human Activity Recognition (HAR) has recently received remarkable attention\nin numerous applications such as assisted living and remote monitoring.\nExisting solutions based on sensors and vision technologies have obtained\nachievements but still suffering from considerable limitations in the\nenvironmental requirement. Wireless signals like WiFi-based sensing have\nemerged as a new paradigm since it is convenient and not restricted in the\nenvironment. In this paper, a new WiFi-based and video-based neural network\n(WiNN) is proposed to improve the robustness of activity recognition where the\nsynchronized video serves as the supplement for the wireless data. Moreover, a\nwireless-vision benchmark (WiVi) is collected for 9 class actions recognition\nin three different visual conditions, including the scenes without occlusion,\nwith partial occlusion, and with full occlusion. Both machine learning methods\n- support vector machine (SVM) as well as deep learning methods are used for\nthe accuracy verification of the data set. Our results show that WiVi data set\nsatisfies the primary demand and all three branches in the proposed pipeline\nkeep more than $80\\%$ of activity recognition accuracy over multiple action\nsegmentation from 1s to 3s. In particular, WiNN is the most robust method in\nterms of all the actions on three action segmentation compared to the others.",
"title": "A Wireless-Vision Dataset for Privacy Preserving Human Activity Recognition",
"url": "http://arxiv.org/abs/2205.11962v1"
} | null | null | new_dataset | admin | null | false | null | 536560d9-c1f6-4c9d-8626-465be39d4d01 | null | Validated | 2023-10-04 15:19:51.886206 | {
"text_length": 1482
} | 0new_dataset
|
TITLE: Multi Visual Modality Fall Detection Dataset
ABSTRACT: Falls are one of the leading cause of injury-related deaths among the elderly
worldwide. Effective detection of falls can reduce the risk of complications
and injuries. Fall detection can be performed using wearable devices or ambient
sensors; these methods may struggle with user compliance issues or false
alarms. Video cameras provide a passive alternative; however, regular RGB
cameras are impacted by changing lighting conditions and privacy concerns. From
a machine learning perspective, developing an effective fall detection system
is challenging because of the rarity and variability of falls. Many existing
fall detection datasets lack important real-world considerations, such as
varied lighting, continuous activities of daily living (ADLs), and camera
placement. The lack of these considerations makes it difficult to develop
predictive models that can operate effectively in the real world. To address
these limitations, we introduce a novel multi-modality dataset (MUVIM) that
contains four visual modalities: infra-red, depth, RGB and thermal cameras.
These modalities offer benefits such as obfuscated facial features and improved
performance in low-light conditions. We formulated fall detection as an anomaly
detection problem, in which a customized spatio-temporal convolutional
autoencoder was trained only on ADLs so that a fall would increase the
reconstruction error. Our results showed that infra-red cameras provided the
highest level of performance (AUC ROC=0.94), followed by thermal (AUC
ROC=0.87), depth (AUC ROC=0.86) and RGB (AUC ROC=0.83). This research provides
a unique opportunity to analyze the utility of camera modalities in detecting
falls in a home setting while balancing performance, passiveness, and privacy. | {
"abstract": "Falls are one of the leading cause of injury-related deaths among the elderly\nworldwide. Effective detection of falls can reduce the risk of complications\nand injuries. Fall detection can be performed using wearable devices or ambient\nsensors; these methods may struggle with user compliance issues or false\nalarms. Video cameras provide a passive alternative; however, regular RGB\ncameras are impacted by changing lighting conditions and privacy concerns. From\na machine learning perspective, developing an effective fall detection system\nis challenging because of the rarity and variability of falls. Many existing\nfall detection datasets lack important real-world considerations, such as\nvaried lighting, continuous activities of daily living (ADLs), and camera\nplacement. The lack of these considerations makes it difficult to develop\npredictive models that can operate effectively in the real world. To address\nthese limitations, we introduce a novel multi-modality dataset (MUVIM) that\ncontains four visual modalities: infra-red, depth, RGB and thermal cameras.\nThese modalities offer benefits such as obfuscated facial features and improved\nperformance in low-light conditions. We formulated fall detection as an anomaly\ndetection problem, in which a customized spatio-temporal convolutional\nautoencoder was trained only on ADLs so that a fall would increase the\nreconstruction error. Our results showed that infra-red cameras provided the\nhighest level of performance (AUC ROC=0.94), followed by thermal (AUC\nROC=0.87), depth (AUC ROC=0.86) and RGB (AUC ROC=0.83). This research provides\na unique opportunity to analyze the utility of camera modalities in detecting\nfalls in a home setting while balancing performance, passiveness, and privacy.",
"title": "Multi Visual Modality Fall Detection Dataset",
"url": "http://arxiv.org/abs/2206.12740v1"
} | null | null | new_dataset | admin | null | false | null | 9e11cce3-f6fa-437c-9be7-a30019f60463 | null | Validated | 2023-10-04 15:19:51.885583 | {
"text_length": 1831
} | 0new_dataset
|
TITLE: Examining Uniqueness and Permanence of the WAY EEG GAL dataset toward User Authentication
ABSTRACT: This study evaluates the discriminating capacity (uniqueness) of the EEG data
from the WAY EEG GAL public dataset to authenticate individuals against one
another as well as its permanence. In addition to the EEG data, Luciw et al.
provide EMG (Electromyography), and kinematics data for engineers and
researchers to utilize WAY EEG GAL for further studies. However, evaluating the
EMG and kinematics data is outside the scope of this study. The goal of the
state-of-the-art is to determine whether EEG data can be utilized to control
prosthetic devices. On the other hand, this study aims to evaluate the
separability of individuals through EEG data to perform user authentication. A
feature importance algorithm is utilized to select the best features for each
user to authenticate them against all others. The authentication platform
implemented for this study is based on Machine Learning models/classifiers. As
an initial test, two pilot studies are performed using Linear Discriminant
Analysis (LDA) and Support Vector Machine (SVM) to observe the learning trends
of the models by multi-labeling the EEG dataset. Utilizing kNN first as the
classifier for user authentication, accuracy around 75% is observed. Thereafter
to improve the performance both linear and non-linear SVMs are used to perform
classification. The overall average accuracies of 85.18% and 86.92% are
achieved using linear and non-linear SVMs respectively. In addition to
accuracy, F1 scores are also calculated. The overall average F1 score of 87.51%
and 88.94% are achieved for linear and non-linear SVMs respectively. Beyond the
overall performance, high performing individuals with 95.3% accuracy (95.3% F1
score) using linear SVM and 97.4% accuracy (97.3% F1 score) using non-linear
SVM are also observed. | {
"abstract": "This study evaluates the discriminating capacity (uniqueness) of the EEG data\nfrom the WAY EEG GAL public dataset to authenticate individuals against one\nanother as well as its permanence. In addition to the EEG data, Luciw et al.\nprovide EMG (Electromyography), and kinematics data for engineers and\nresearchers to utilize WAY EEG GAL for further studies. However, evaluating the\nEMG and kinematics data is outside the scope of this study. The goal of the\nstate-of-the-art is to determine whether EEG data can be utilized to control\nprosthetic devices. On the other hand, this study aims to evaluate the\nseparability of individuals through EEG data to perform user authentication. A\nfeature importance algorithm is utilized to select the best features for each\nuser to authenticate them against all others. The authentication platform\nimplemented for this study is based on Machine Learning models/classifiers. As\nan initial test, two pilot studies are performed using Linear Discriminant\nAnalysis (LDA) and Support Vector Machine (SVM) to observe the learning trends\nof the models by multi-labeling the EEG dataset. Utilizing kNN first as the\nclassifier for user authentication, accuracy around 75% is observed. Thereafter\nto improve the performance both linear and non-linear SVMs are used to perform\nclassification. The overall average accuracies of 85.18% and 86.92% are\nachieved using linear and non-linear SVMs respectively. In addition to\naccuracy, F1 scores are also calculated. The overall average F1 score of 87.51%\nand 88.94% are achieved for linear and non-linear SVMs respectively. Beyond the\noverall performance, high performing individuals with 95.3% accuracy (95.3% F1\nscore) using linear SVM and 97.4% accuracy (97.3% F1 score) using non-linear\nSVM are also observed.",
"title": "Examining Uniqueness and Permanence of the WAY EEG GAL dataset toward User Authentication",
"url": "http://arxiv.org/abs/2209.04802v1"
} | null | null | no_new_dataset | admin | null | false | null | bda01d07-ebc9-4f0c-bcfc-c2b2b271318e | null | Validated | 2023-10-04 15:19:51.884090 | {
"text_length": 1909
} | 1no_new_dataset
|
TITLE: A curated dataset for data-driven turbulence modelling
ABSTRACT: The recent surge in machine learning augmented turbulence modelling is a
promising approach for addressing the limitations of Reynolds-averaged
Navier-Stokes (RANS) models. This work presents the development of the first
open-source dataset, curated and structured for immediate use in machine
learning augmented turbulence closure modelling. The dataset features a variety
of RANS simulations with matching direct numerical simulation (DNS) and
large-eddy simulation (LES) data. Four turbulence models are selected to form
the initial dataset: $k$-$\varepsilon$, $k$-$\varepsilon$-$\phi_t$-$f$,
$k$-$\omega$, and $k$-$\omega$ SST. The dataset consists of 29 cases per
turbulence model, for several parametrically sweeping reference DNS/LES cases:
periodic hills, square duct, parametric bumps, converging-diverging channel,
and a curved backward-facing step. At each of the 895,640 points, various RANS
features with DNS/LES labels are available. The feature set includes quantities
used in current state-of-the-art models, and additional fields which enable the
generation of new feature sets. The dataset reduces effort required to train,
test, and benchmark new models. The dataset is available at
https://doi.org/10.34740/kaggle/dsv/2044393 . | {
"abstract": "The recent surge in machine learning augmented turbulence modelling is a\npromising approach for addressing the limitations of Reynolds-averaged\nNavier-Stokes (RANS) models. This work presents the development of the first\nopen-source dataset, curated and structured for immediate use in machine\nlearning augmented turbulence closure modelling. The dataset features a variety\nof RANS simulations with matching direct numerical simulation (DNS) and\nlarge-eddy simulation (LES) data. Four turbulence models are selected to form\nthe initial dataset: $k$-$\\varepsilon$, $k$-$\\varepsilon$-$\\phi_t$-$f$,\n$k$-$\\omega$, and $k$-$\\omega$ SST. The dataset consists of 29 cases per\nturbulence model, for several parametrically sweeping reference DNS/LES cases:\nperiodic hills, square duct, parametric bumps, converging-diverging channel,\nand a curved backward-facing step. At each of the 895,640 points, various RANS\nfeatures with DNS/LES labels are available. The feature set includes quantities\nused in current state-of-the-art models, and additional fields which enable the\ngeneration of new feature sets. The dataset reduces effort required to train,\ntest, and benchmark new models. The dataset is available at\nhttps://doi.org/10.34740/kaggle/dsv/2044393 .",
"title": "A curated dataset for data-driven turbulence modelling",
"url": "http://arxiv.org/abs/2103.11515v1"
} | null | null | new_dataset | admin | null | false | null | 01c0c89e-e0ac-4fb0-a0c7-33d0c42c3e36 | null | Validated | 2023-10-04 15:19:51.895336 | {
"text_length": 1336
} | 0new_dataset
|
TITLE: A hands-on gaze on HTTP/3 security through the lens of HTTP/2 and a public dataset
ABSTRACT: Following QUIC protocol ratification on May 2021, the third major version of
the Hypertext Transfer Protocol, namely HTTP/3, was published around one year
later in RFC 9114. In light of these consequential advancements, the current
work aspires to provide a full-blown coverage of the following issues, which to
our knowledge have received feeble or no attention in the literature so far.
First, we provide a complete review of attacks against HTTP/2, and elaborate on
if and in which way they can be migrated to HTTP/3. Second, through the
creation of a testbed comprising the at present six most popular HTTP/3-enabled
servers, we examine the effectiveness of a quartet of attacks, either stemming
directly from the HTTP/2 relevant literature or being entirely new. This
scrutiny led to the assignment of at least one CVE ID with a critical base
score by MITRE. No less important, by capitalizing on a realistic, abundant in
devices testbed, we compiled a voluminous, labeled corpus containing traces of
ten diverse attacks against HTTP and QUIC services. An initial evaluation of
the dataset mainly by means of machine learning techniques is included as well.
Given that the 30 GB dataset is made available in both pcap and CSV formats,
forthcoming research can easily take advantage of any subset of features,
contingent upon the specific network topology and configuration. | {
"abstract": "Following QUIC protocol ratification on May 2021, the third major version of\nthe Hypertext Transfer Protocol, namely HTTP/3, was published around one year\nlater in RFC 9114. In light of these consequential advancements, the current\nwork aspires to provide a full-blown coverage of the following issues, which to\nour knowledge have received feeble or no attention in the literature so far.\nFirst, we provide a complete review of attacks against HTTP/2, and elaborate on\nif and in which way they can be migrated to HTTP/3. Second, through the\ncreation of a testbed comprising the at present six most popular HTTP/3-enabled\nservers, we examine the effectiveness of a quartet of attacks, either stemming\ndirectly from the HTTP/2 relevant literature or being entirely new. This\nscrutiny led to the assignment of at least one CVE ID with a critical base\nscore by MITRE. No less important, by capitalizing on a realistic, abundant in\ndevices testbed, we compiled a voluminous, labeled corpus containing traces of\nten diverse attacks against HTTP and QUIC services. An initial evaluation of\nthe dataset mainly by means of machine learning techniques is included as well.\nGiven that the 30 GB dataset is made available in both pcap and CSV formats,\nforthcoming research can easily take advantage of any subset of features,\ncontingent upon the specific network topology and configuration.",
"title": "A hands-on gaze on HTTP/3 security through the lens of HTTP/2 and a public dataset",
"url": "http://arxiv.org/abs/2208.06722v2"
} | null | null | new_dataset | admin | null | false | null | 21e2b6c8-26fb-4355-b406-0515c5b2df27 | null | Validated | 2023-10-04 15:19:51.884840 | {
"text_length": 1495
} | 0new_dataset
|
TITLE: Quantum Transfer Learning for Real-World, Small, and High-Dimensional Datasets
ABSTRACT: Quantum machine learning (QML) networks promise to have some computational
(or quantum) advantage for classifying supervised datasets (e.g., satellite
images) over some conventional deep learning (DL) techniques due to their
expressive power via their local effective dimension. There are, however, two
main challenges regardless of the promised quantum advantage: 1) Currently
available quantum bits (qubits) are very small in number, while real-world
datasets are characterized by hundreds of high-dimensional elements (i.e.,
features). Additionally, there is not a single unified approach for embedding
real-world high-dimensional datasets in a limited number of qubits. 2) Some
real-world datasets are too small for training intricate QML networks. Hence,
to tackle these two challenges for benchmarking and validating QML networks on
real-world, small, and high-dimensional datasets in one-go, we employ quantum
transfer learning composed of a multi-qubit QML network, and a very deep
convolutional network (a with VGG16 architecture) extracting informative
features from any small, high-dimensional dataset. We use real-amplitude and
strongly-entangling N-layer QML networks with and without data re-uploading
layers as a multi-qubit QML network, and evaluate their expressive power
quantified by using their local effective dimension; the lower the local
effective dimension of a QML network, the better its performance on unseen
data. Our numerical results show that the strongly-entangling N-layer QML
network has a lower local effective dimension than the real-amplitude QML
network and outperforms it on the hard-to-classify three-class labelling
problem. In addition, quantum transfer learning helps tackle the two challenges
mentioned above for benchmarking and validating QML networks on real-world,
small, and high-dimensional datasets. | {
"abstract": "Quantum machine learning (QML) networks promise to have some computational\n(or quantum) advantage for classifying supervised datasets (e.g., satellite\nimages) over some conventional deep learning (DL) techniques due to their\nexpressive power via their local effective dimension. There are, however, two\nmain challenges regardless of the promised quantum advantage: 1) Currently\navailable quantum bits (qubits) are very small in number, while real-world\ndatasets are characterized by hundreds of high-dimensional elements (i.e.,\nfeatures). Additionally, there is not a single unified approach for embedding\nreal-world high-dimensional datasets in a limited number of qubits. 2) Some\nreal-world datasets are too small for training intricate QML networks. Hence,\nto tackle these two challenges for benchmarking and validating QML networks on\nreal-world, small, and high-dimensional datasets in one-go, we employ quantum\ntransfer learning composed of a multi-qubit QML network, and a very deep\nconvolutional network (a with VGG16 architecture) extracting informative\nfeatures from any small, high-dimensional dataset. We use real-amplitude and\nstrongly-entangling N-layer QML networks with and without data re-uploading\nlayers as a multi-qubit QML network, and evaluate their expressive power\nquantified by using their local effective dimension; the lower the local\neffective dimension of a QML network, the better its performance on unseen\ndata. Our numerical results show that the strongly-entangling N-layer QML\nnetwork has a lower local effective dimension than the real-amplitude QML\nnetwork and outperforms it on the hard-to-classify three-class labelling\nproblem. In addition, quantum transfer learning helps tackle the two challenges\nmentioned above for benchmarking and validating QML networks on real-world,\nsmall, and high-dimensional datasets.",
"title": "Quantum Transfer Learning for Real-World, Small, and High-Dimensional Datasets",
"url": "http://arxiv.org/abs/2209.07799v4"
} | null | null | no_new_dataset | admin | null | false | null | c751cb1f-cd90-46e8-8490-989b40bf0b76 | null | Validated | 2023-10-04 15:19:51.884007 | {
"text_length": 1964
} | 1no_new_dataset
|
TITLE: A Sentence Cloze Dataset for Chinese Machine Reading Comprehension
ABSTRACT: Owing to the continuous efforts by the Chinese NLP community, more and more
Chinese machine reading comprehension datasets become available. To add
diversity in this area, in this paper, we propose a new task called Sentence
Cloze-style Machine Reading Comprehension (SC-MRC). The proposed task aims to
fill the right candidate sentence into the passage that has several blanks. We
built a Chinese dataset called CMRC 2019 to evaluate the difficulty of the
SC-MRC task. Moreover, to add more difficulties, we also made fake candidates
that are similar to the correct ones, which requires the machine to judge their
correctness in the context. The proposed dataset contains over 100K blanks
(questions) within over 10K passages, which was originated from Chinese
narrative stories. To evaluate the dataset, we implement several baseline
systems based on the pre-trained models, and the results show that the
state-of-the-art model still underperforms human performance by a large margin.
We release the dataset and baseline system to further facilitate our community.
Resources available through https://github.com/ymcui/cmrc2019 | {
"abstract": "Owing to the continuous efforts by the Chinese NLP community, more and more\nChinese machine reading comprehension datasets become available. To add\ndiversity in this area, in this paper, we propose a new task called Sentence\nCloze-style Machine Reading Comprehension (SC-MRC). The proposed task aims to\nfill the right candidate sentence into the passage that has several blanks. We\nbuilt a Chinese dataset called CMRC 2019 to evaluate the difficulty of the\nSC-MRC task. Moreover, to add more difficulties, we also made fake candidates\nthat are similar to the correct ones, which requires the machine to judge their\ncorrectness in the context. The proposed dataset contains over 100K blanks\n(questions) within over 10K passages, which was originated from Chinese\nnarrative stories. To evaluate the dataset, we implement several baseline\nsystems based on the pre-trained models, and the results show that the\nstate-of-the-art model still underperforms human performance by a large margin.\nWe release the dataset and baseline system to further facilitate our community.\nResources available through https://github.com/ymcui/cmrc2019",
"title": "A Sentence Cloze Dataset for Chinese Machine Reading Comprehension",
"url": "http://arxiv.org/abs/2004.03116v2"
} | null | null | new_dataset | admin | null | false | null | 3dd4fa79-af91-4917-bdb2-2fbd39c41dc5 | null | Validated | 2023-10-04 15:19:51.901179 | {
"text_length": 1229
} | 0new_dataset
|
TITLE: ML4CO-KIDA: Knowledge Inheritance in Dataset Aggregation
ABSTRACT: The Machine Learning for Combinatorial Optimization (ML4CO) NeurIPS 2021
competition aims to improve state-of-the-art combinatorial optimization solvers
by replacing key heuristic components with machine learning models. On the dual
task, we design models to make branching decisions to promote the dual bound
increase faster. We propose a knowledge inheritance method to generalize
knowledge of different models from the dataset aggregation process, named KIDA.
Our improvement overcomes some defects of the baseline
graph-neural-networks-based methods. Further, we won the
$1$\textsuperscript{st} Place on the dual task. We hope this report can provide
useful experience for developers and researchers. The code is available at
https://github.com/megvii-research/NeurIPS2021-ML4CO-KIDA. | {
"abstract": "The Machine Learning for Combinatorial Optimization (ML4CO) NeurIPS 2021\ncompetition aims to improve state-of-the-art combinatorial optimization solvers\nby replacing key heuristic components with machine learning models. On the dual\ntask, we design models to make branching decisions to promote the dual bound\nincrease faster. We propose a knowledge inheritance method to generalize\nknowledge of different models from the dataset aggregation process, named KIDA.\nOur improvement overcomes some defects of the baseline\ngraph-neural-networks-based methods. Further, we won the\n$1$\\textsuperscript{st} Place on the dual task. We hope this report can provide\nuseful experience for developers and researchers. The code is available at\nhttps://github.com/megvii-research/NeurIPS2021-ML4CO-KIDA.",
"title": "ML4CO-KIDA: Knowledge Inheritance in Dataset Aggregation",
"url": "http://arxiv.org/abs/2201.10328v3"
} | null | null | no_new_dataset | admin | null | false | null | b6ce548a-2c17-491d-b3e4-b0fd6e353911 | null | Validated | 2023-10-04 15:19:51.888653 | {
"text_length": 879
} | 1no_new_dataset
|
TITLE: A Retail Product Categorisation Dataset
ABSTRACT: Most eCommerce applications, like web-shops have millions of products. In
this context, the identification of similar products is a common sub-task,
which can be utilized in the implementation of recommendation systems, product
search engines and internal supply logistics. Providing this data set, our goal
is to boost the evaluation of machine learning methods for the prediction of
the category of the retail products from tuples of images and descriptions. | {
"abstract": "Most eCommerce applications, like web-shops have millions of products. In\nthis context, the identification of similar products is a common sub-task,\nwhich can be utilized in the implementation of recommendation systems, product\nsearch engines and internal supply logistics. Providing this data set, our goal\nis to boost the evaluation of machine learning methods for the prediction of\nthe category of the retail products from tuples of images and descriptions.",
"title": "A Retail Product Categorisation Dataset",
"url": "http://arxiv.org/abs/2103.13864v2"
} | null | null | new_dataset | admin | null | false | null | 4d1da674-b4e8-4b2d-89e4-44197dc59c38 | null | Validated | 2023-10-04 15:19:51.895238 | {
"text_length": 534
} | 0new_dataset
|
TITLE: TabNAS: Rejection Sampling for Neural Architecture Search on Tabular Datasets
ABSTRACT: The best neural architecture for a given machine learning problem depends on
many factors: not only the complexity and structure of the dataset, but also on
resource constraints including latency, compute, energy consumption, etc.
Neural architecture search (NAS) for tabular datasets is an important but
under-explored problem. Previous NAS algorithms designed for image search
spaces incorporate resource constraints directly into the reinforcement
learning (RL) rewards. However, for NAS on tabular datasets, this protocol
often discovers suboptimal architectures. This paper develops TabNAS, a new and
more effective approach to handle resource constraints in tabular NAS using an
RL controller motivated by the idea of rejection sampling. TabNAS immediately
discards any architecture that violates the resource constraints without
training or learning from that architecture. TabNAS uses a Monte-Carlo-based
correction to the RL policy gradient update to account for this extra filtering
step. Results on several tabular datasets demonstrate the superiority of TabNAS
over previous reward-shaping methods: it finds better models that obey the
constraints. | {
"abstract": "The best neural architecture for a given machine learning problem depends on\nmany factors: not only the complexity and structure of the dataset, but also on\nresource constraints including latency, compute, energy consumption, etc.\nNeural architecture search (NAS) for tabular datasets is an important but\nunder-explored problem. Previous NAS algorithms designed for image search\nspaces incorporate resource constraints directly into the reinforcement\nlearning (RL) rewards. However, for NAS on tabular datasets, this protocol\noften discovers suboptimal architectures. This paper develops TabNAS, a new and\nmore effective approach to handle resource constraints in tabular NAS using an\nRL controller motivated by the idea of rejection sampling. TabNAS immediately\ndiscards any architecture that violates the resource constraints without\ntraining or learning from that architecture. TabNAS uses a Monte-Carlo-based\ncorrection to the RL policy gradient update to account for this extra filtering\nstep. Results on several tabular datasets demonstrate the superiority of TabNAS\nover previous reward-shaping methods: it finds better models that obey the\nconstraints.",
"title": "TabNAS: Rejection Sampling for Neural Architecture Search on Tabular Datasets",
"url": "http://arxiv.org/abs/2204.07615v4"
} | null | null | no_new_dataset | admin | null | false | null | cbdfb7e9-ecd1-425b-8196-da26483e8cae | null | Validated | 2023-10-04 15:19:51.887039 | {
"text_length": 1272
} | 1no_new_dataset
|
TITLE: When More is Less: Incorporating Additional Datasets Can Hurt Performance By Introducing Spurious Correlations
ABSTRACT: In machine learning, incorporating more data is often seen as a reliable
strategy for improving model performance; this work challenges that notion by
demonstrating that the addition of external datasets in many cases can hurt the
resulting model's performance. In a large-scale empirical study across
combinations of four different open-source chest x-ray datasets and 9 different
labels, we demonstrate that in 43% of settings, a model trained on data from
two hospitals has poorer worst group accuracy over both hospitals than a model
trained on just a single hospital's data. This surprising result occurs even
though the added hospital makes the training distribution more similar to the
test distribution. We explain that this phenomenon arises from the spurious
correlation that emerges between the disease and hospital, due to
hospital-specific image artifacts. We highlight the trade-off one encounters
when training on multiple datasets, between the obvious benefit of additional
data and insidious cost of the introduced spurious correlation. In some cases,
balancing the dataset can remove the spurious correlation and improve
performance, but it is not always an effective strategy. We contextualize our
results within the literature on spurious correlations to help explain these
outcomes. Our experiments underscore the importance of exercising caution when
selecting training data for machine learning models, especially in settings
where there is a risk of spurious correlations such as with medical imaging.
The risks outlined highlight the need for careful data selection and model
evaluation in future research and practice. | {
"abstract": "In machine learning, incorporating more data is often seen as a reliable\nstrategy for improving model performance; this work challenges that notion by\ndemonstrating that the addition of external datasets in many cases can hurt the\nresulting model's performance. In a large-scale empirical study across\ncombinations of four different open-source chest x-ray datasets and 9 different\nlabels, we demonstrate that in 43% of settings, a model trained on data from\ntwo hospitals has poorer worst group accuracy over both hospitals than a model\ntrained on just a single hospital's data. This surprising result occurs even\nthough the added hospital makes the training distribution more similar to the\ntest distribution. We explain that this phenomenon arises from the spurious\ncorrelation that emerges between the disease and hospital, due to\nhospital-specific image artifacts. We highlight the trade-off one encounters\nwhen training on multiple datasets, between the obvious benefit of additional\ndata and insidious cost of the introduced spurious correlation. In some cases,\nbalancing the dataset can remove the spurious correlation and improve\nperformance, but it is not always an effective strategy. We contextualize our\nresults within the literature on spurious correlations to help explain these\noutcomes. Our experiments underscore the importance of exercising caution when\nselecting training data for machine learning models, especially in settings\nwhere there is a risk of spurious correlations such as with medical imaging.\nThe risks outlined highlight the need for careful data selection and model\nevaluation in future research and practice.",
"title": "When More is Less: Incorporating Additional Datasets Can Hurt Performance By Introducing Spurious Correlations",
"url": "http://arxiv.org/abs/2308.04431v1"
} | null | null | no_new_dataset | admin | null | false | null | 8fe7555f-fb7d-4a70-bb98-dcc9e18168a6 | null | Validated | 2023-10-04 15:19:51.864300 | {
"text_length": 1789
} | 1no_new_dataset
|
TITLE: A Case for Dataset Specific Profiling
ABSTRACT: Data-driven science is an emerging paradigm where scientific discoveries
depend on the execution of computational AI models against rich,
discipline-specific datasets. With modern machine learning frameworks, anyone
can develop and execute computational models that reveal concepts hidden in the
data that could enable scientific applications. For important and widely used
datasets, computing the performance of every computational model that can run
against a dataset is cost prohibitive in terms of cloud resources. Benchmarking
approaches used in practice use representative datasets to infer performance
without actually executing models. While practicable, these approaches limit
extensive dataset profiling to a few datasets and introduce bias that favors
models suited for representative datasets. As a result, each dataset's unique
characteristics are left unexplored and subpar models are selected based on
inference from generalized datasets. This necessitates a new paradigm that
introduces dataset profiling into the model selection process. To demonstrate
the need for dataset-specific profiling, we answer two questions:(1) Can
scientific datasets significantly permute the rank order of computational
models compared to widely used representative datasets? (2) If so, could
lightweight model execution improve benchmarking accuracy? Taken together, the
answers to these questions lay the foundation for a new dataset-aware
benchmarking paradigm. | {
"abstract": "Data-driven science is an emerging paradigm where scientific discoveries\ndepend on the execution of computational AI models against rich,\ndiscipline-specific datasets. With modern machine learning frameworks, anyone\ncan develop and execute computational models that reveal concepts hidden in the\ndata that could enable scientific applications. For important and widely used\ndatasets, computing the performance of every computational model that can run\nagainst a dataset is cost prohibitive in terms of cloud resources. Benchmarking\napproaches used in practice use representative datasets to infer performance\nwithout actually executing models. While practicable, these approaches limit\nextensive dataset profiling to a few datasets and introduce bias that favors\nmodels suited for representative datasets. As a result, each dataset's unique\ncharacteristics are left unexplored and subpar models are selected based on\ninference from generalized datasets. This necessitates a new paradigm that\nintroduces dataset profiling into the model selection process. To demonstrate\nthe need for dataset-specific profiling, we answer two questions:(1) Can\nscientific datasets significantly permute the rank order of computational\nmodels compared to widely used representative datasets? (2) If so, could\nlightweight model execution improve benchmarking accuracy? Taken together, the\nanswers to these questions lay the foundation for a new dataset-aware\nbenchmarking paradigm.",
"title": "A Case for Dataset Specific Profiling",
"url": "http://arxiv.org/abs/2208.03315v1"
} | null | null | no_new_dataset | admin | null | false | null | 7df67111-f805-406e-bf38-bae6b04c1dca | null | Validated | 2023-10-04 15:19:51.884946 | {
"text_length": 1533
} | 1no_new_dataset
|