Data labeling / ethics and bias mitigationdataset creation and annotation ethicsParticipation documented

Rehumanized Crowdsourcing

Syracuse, United States

The framework explicitly re-centers human labelers in ML annotation workflows to address bias and ethics.

human-centered designlabor transparencyethics in annotationdata collectionlabelingquality control
Overview
Region
North America
Lead organization
Syracuse University with Figure Eight
Organization type
university + industry partnership
Technology group
data labeling workflows for machine learning
Activity status
published_case
Start year
2019
Last updated
Mar 25, 2026
Details
Participation mode
crowdworker-centered labeling framework design
Participation group
Community In The Loop
Technology description
data labeling workflows for machine learning
Funding
Not documented
Region of activity
national

Community input

Corrections, annotations, and private disclosure requests.

Rehumanized Crowdsourcing has 0 published community items.

Published corrections

Loading moderation history...

Published annotations

Loading annotations...

Submit a correction

Flag a metadata problem or submit a dispute for Rehumanized Crowdsourcing.

Submit an annotation

Add context without editing the record itself.

Request restricted disclosure

These requests are private and reviewed for later releases.

Related records

Algorithmic governance / public servicesNorth America
AI Failure Cards for Homeless Services

Pittsburgh, United States

Open record
Participatory auditing / LLM safetyNorth America
AI Red Teaming Challenge

Las Vegas, United States

Open record
Governance / auditingNorth America
Algorithmic Equity Toolkit

Seattle, United States

Open record
Social media / contextual NLPNorth America
Artificial Intelligence and Inclusion: Formerly Gang-Involved Youth as Domain Experts

New York City, United States

Open record