PDF Viewer

BUILDER'S SANDBOX

Build This Paper

Use an AI coding agent to implement this research.

OpenAI Codex
OpenAI CodexAI Agent

Lightweight coding agent in your terminal.

Claude Code
Claude CodeAI Agent

Agentic coding tool for terminal workflows.

AntiGravity IDE
AntiGravity IDEScaffolding

AI agent mindset installer and workflow scaffolder.

Cursor
CursorIDE

AI-first code editor built on VS Code.

VS Code
VS CodeIDE

Free, open-source editor by Microsoft.

MVP Investment

$9K - $12K
6-10 weeks
Engineering
$8,000
Cloud Hosting
$240
SaaS Stack
$300
Domain & Legal
$100

6mo ROI

2-4x

3yr ROI

10-20x

Lightweight AI tools can reach profitability quickly. At $500/mo average contract, 20 customers = $10K MRR by 6mo, 200+ by 3yr.

Talent Scout

L

Laura De Grazia

CLiC–Language and Computing Center, University of Barcelona

D

Danae Sánchez Villegas

Department of Computer Science, University of Copenhagen

D

Desmond Elliott

Department of Computer Science, University of Copenhagen

M

Mireia Farrús

CLiC–Language and Computing Center, University of Barcelona

Find Similar Experts

Social experts on LinkedIn & GitHub

References (55)

[1]
A Literature Survey on Multimodal and Multilingual Sexism Detection
2025Xuan Luo, Bin Liang et al.
[2]
Spanish is not just one: A dataset of Spanish dialect recognition for LLMs
2025Gonzalo Martínez, Marina Mayor-Rocher et al.
[3]
Social media use and roles of self-objectification, self-compassion and body image concerns: a systematic review
2025Elisa Sarda, Claire El-Jor et al.
[4]
HateClipSeg: A Segment-Level Annotated Dataset for Fine-Grained Hate Video Detection
2025Han Wang, Zhuoran Wang et al.
[5]
Beyond Binary Moderation: Identifying Fine-Grained Sexist and Misogynistic Behavior on GitHub with Large Language Models
2025Tanni Dev, Sayma Sultana et al.
[6]
Gendered disinformation as violence: A new analytical agenda
2025Marília Gehrke, Eedan R. Amit-Danhi
[7]
MuSeD: A Multimodal Spanish Dataset for Sexism Detection in Social Media Videos
2025Laura De Grazia, Pol Pastells et al.
[8]
Salamandra Technical Report
2025Aitor Gonzalez-Agirre, Marc Pàmies et al.
[9]
Interpretable Sexism Detection with Explainable Transformers
2025Shamima Rayhana, Md Shajalal et al.
[10]
Large Language Models with Reinforcement Learning from Human Feedback Approach for Enhancing Explainable Sexism Detection
2025A. Samani, Tianhao Wang et al.
[11]
Qwen2.5 Technical Report
2024Qwen An Yang, Baosong Yang et al.
[12]
Applying a Gender Lens to the Study of Misinformation and Disinformation
2024Gergana Tzvetkova
[13]
The Llama 3 Herd of Models
2024Abhimanyu Dubey, Abhinav Jauhri et al.
[14]
MultiHateClip: A Multilingual Benchmark Dataset for Hateful Video Detection on YouTube and Bilibili
2024Han Wang, Tan Rui Yang et al.
[15]
MIMIC: Misogyny Identification in Multimodal Internet Content in Hindi-English Code-Mixed Language
2024Aakash Singh, Deepawali Sharma et al.
[16]
Shorts vs. Regular Videos on YouTube: A Comparative Analysis of User Engagement and Content Creation Trends
2024Caroline Violot, Tuğrulcan Elmas et al.
[17]
EXIST 2024: sEXism Identification in Social neTworks and Memes
2024Laura Plaza, Jorge Carrillo-de-Albornoz et al.
[18]
From Laughter to Inequality: Annotated Dataset for Misogyny Detection in Tamil and Malayalam Memes
2024Rahul Ponnusamy, Kathiravan Pannerselvam et al.
[19]
Detection of homophobia and transphobia in YouTube comments
2023Bharathi Raja Chakravarthi
[20]
Evaluating GPT-3 Generated Explanations for Hateful Content Moderation
2023H. Wang, Ming Shan Hee et al.

Showing 20 of 55 references

Founder's Pitch

"Develop a tool to fine-tune social media platforms' sexist content detection using FineMuSe, a multimodal dataset for fine-grained analysis."

Social Media AnalysisScore: 4View PDF ↗

Commercial Viability Breakdown

0-10 scale

High Potential

2/4 signals

5

Quick Build

4/4 signals

10

Series A Potential

2/4 signals

5

Sources used for this analysis

arXiv Paper

Full-text PDF analysis of the research paper

GitHub Repository

Code availability, stars, and contributor activity

Citation Network

Semantic Scholar citations and co-citation patterns

Community Predictions

Crowd-sourced unicorn probability assessments

Analysis model: GPT-4o · Last scored: 2/17/2026

🔭 Research Neighborhood

Generating constellation...

~3-8 seconds

Why It Matters

This research breaks the limitations of binary sexism classification, allowing for more nuanced detection which is essential in tackling the complex manifestations of sexist content online.

Product Angle

Leverage the FineMuSe dataset to develop a commercial API or software tool that integrates with social media platforms, enhancing their content moderation capabilities with fine-grained sexism detection.

Disruption

This approach could replace basic moderation tools that currently rely on binary metrics, offering a detailed understanding of content to better align with community standards.

Product Opportunity

The market for social media content moderation is growing as platforms seek to improve user experience by filtering harmful content. Companies managing social media sites pay for advanced moderation tools.

Use Case Idea

Create a content moderation API for social media platforms to automatically detect and classify sexism with fine-grained distinctions, improving content filtering systems.

Science

The paper introduces FineMuSe, a multimodal dataset with binary and fine-grained annotations of sexism in Spanish social media videos. It evaluates various LLMs on this dataset to detect fine-grained sexism using text, audio, and video modalities.

Method & Eval

The study utilized multimodal LLMs to annotate and classify sexist content in social media videos, comparing machine learning outputs with human annotations to assess accuracy in detecting subtle forms of sexism.

Caveats

The dataset and models may not generalize across different languages or cultural contexts beyond Spanish content, potentially limiting the tool's broader application.

Author Intelligence

Laura De Grazia

CLiC–Language and Computing Center, University of Barcelona
lauradegrazia@ub.edu

Danae Sánchez Villegas

Department of Computer Science, University of Copenhagen
davi@di.ku.dk

Desmond Elliott

Department of Computer Science, University of Copenhagen
de@di.ku.dk

Mireia Farrús

CLiC–Language and Computing Center, University of Barcelona
mfarrus@ub.edu

Mariona Taulé

CLiC–Language and Computing Center, University of Barcelona
mtaule@ub.edu