Results

Datasets

This section showcases datasets created within the HYBRIDS project. These datasets are designed to support research on disinformation, abusive language, and public discourse analysis. They are periodically updated and made available to the research community to foster collaboration and advance knowledge in the field.

datasets-Raluca

MetaHate: A Unified Dataset for Hate Speech Detection

MetaHate compiles over 1.2 million social media posts from 36 datasets, providing a unified resource for hate speech detection research. Available in TSV format with binary labels, it supports computational linguistics and social media analysis. Access via Hugging Face, with subsets open and full data requiring agreements.

Software and Models

Here, you will find software tools and computational models developed as part of the HYBRIDS project. These resources include AI-driven models for disinformation detection, NLP applications, and hybrid intelligence systems. The section will be continuously updated to provide access to the latest innovations and contributions from the project.

MetaHateBERT

MetaHateBERT

MetaHateBERT is a fine-tuned BERT model specifically designed to detect hate speech in text. It is based on the bert-base-uncased architecture and has been trained for binary text classification, distinguishing between ‘no hate’ and ‘hate’.

CT BERT

CT-BERT-PRCT

A specialized BERT model fine-tuned to detect Population Replacement Conspiracy Theory (PRCT) content across social media platforms. The model demonstrates good performance in identifying both explicit and implicit PRCT narratives, with decent cross-platform and multilingual generalization capabilities.

Llama-3-8B-Distil-MetaHate

Llama-3-8B-Distil-MetaHate is a distilled version of the Llama 3 architecture, fine-tuned for hate speech detection and explanation. Developed by the Information Retrieval Lab at the University of A Coruña, this model employs Chain-of-Thought reasoning to enhance interpretability in hate speech classification tasks. The model aims to not only detect hate speech but also provide explanations for its classifications.