About These Roles
These remote-first roles support end-to-end AI/ML development, from raw data curation to model scorecards. You will work within structured workflows—labeling, validating, and evaluating datasets and model outputs—so downstream teams can strengthen model reliability and safety. Projects cover NLP, computer vision, and multimodal tasks, including entity tagging, sentiment analysis, summarization grading, prompt evaluation, pairwise preference collection for RLHF, bounding box and polygon annotation, segmentation, quality audits, and policy-aligned content safety labeling.



