Key Responsibilities
Produce high-quality annotations for NLP and computer vision datasets; perform RLHF preference comparisons and prompt evaluation for large language models; run QA evaluation on completed tasks to maintain training data quality; follow annotation guidelines compliance and taxonomy standards; document edge cases and escalate ambiguities; contribute to model performance improvement by spotting systemic issues; support large language model evaluation (instruction following, safety, and factuality); execute content safety labeling across policy tiers; participate in calibration sessions and inter-annotator agreement reviews.



