MemeMind: A Large-Scale Multimodal Dataset with Chain-of-Thought Reasoning for Harmful Meme Detection

Beijing University of Posts and Telecommunications
Beijing Normal University

*Indicates Equal Contribution. #Representing the corresponding author.

You can access our dataset, supplementary materials, project pages, and paper through the following four links.

Abstract

The rapid development of social media has intensified the spread of harmful content. Harmful memes, which integrate both images and text, pose significant challenges for automated detection due to their implicit semantics and complex multimodal interactions. Although existing research has made progress in detection accuracy and interpretability, the lack of a systematic, large-scale, diverse, and highly explainable dataset continues to hinder further advancement in this field. To address this gap, we introduce MemeMind, a novel dataset featuring scientifically rigorous standards, large scale, diversity, bilingual support (Chinese and English), and detailed Chain-of-Thought (CoT) annotations. MemeMind fills critical gaps in current datasets by offering comprehensive labeling and explicit reasoning traces, thereby providing a solid foundation for enhancing harmful meme detection. In addition, we propose an innovative detection framework, MemeGuard, which effectively integrates multimodal information with reasoning process modeling, significantly improving models' ability to understand and identify harmful memes. Extensive experiments conducted on the MemeMind dataset demonstrate that MemeGuard consistently outperforms existing state-of-the-art methods in harmful meme detection tasks.

MemeMind Construction Process

Pipeline

First, guided by principles of public safety, legal compliance, and ethical considerations, we established clear criteria for identifying harmful content in memes. Subsequently, to construct MemeMind, we collected multiple relevant datasets and systematically reclassified the samples based on the defined criteria. Furthermore, inspired by human reasoning processes, we employed Chain-of-Thought (CoT) reasoning to annotate the data by simulating how humans think and reason. To ensure annotation quality, we conducted manual sampling reviews throughout the entire annotation process.

Data Diversity

Harmful Meme Categories

Examples of five categories of harmful memes. Images (a) to (e) are English memes, and (f) to (j) are Chinese memes, each corresponding to a specific type of harmful content: (a, f) Discrimination, (b, g) Offensive, (c, h) Violence, (d, i) Vulgar, (e, j) Dissatisfaction.

Case Studies

Case 1
Case 2
Case 3
Case 4

A selection of case studies showcasing the model’s detection performance and reasoning capabilities across diverse harmful meme scenarios.

Our Method: MemeGuard

Method Diagram

During training, the model first undergoes visual understanding enhancement, followed by detection capability refinement. In the inference stage, the model can output the final detection result and optionally provide a detailed Chain-of-Thought reasoning process.

Comparison

Comparison Results

BibTeX

BibTex Code Here