What You'll Do
Assess English-language text outputs from AI models, identifying issues in sentiment interpretation, factual correctness, and logical reasoning. Apply structured evaluation criteria to score model responses across multiple quality dimensions. Review automated annotations to verify reliability and flag inconsistencies that require correction.
Provide clear, written explanations for each evaluation, ensuring decisions are well-justified and repeatable. Maintain high accuracy and consistency when processing large volumes of data, adhering closely to established guidelines.
Requirements
- Professional-level fluency in English, with advanced reading and writing skills
- Proven ability to analyze text and articulate reasoning clearly in writing
- Experience in data annotation, AI training, or similar technical evaluation work
- Strong attention to detail and commitment to data precision
- Demonstrated critical thinking and logical analysis capabilities
- Reliable consistency in judgment across repetitive tasks
Preferred Qualifications
Background in fields requiring analytical rigor—such as linguistics, journalism, or research—is advantageous. Familiarity with artificial intelligence, large language models, or automated evaluation systems is helpful but not required.
Benefits
This position supports remote work for candidates based in the Philippines. The role contributes directly to the improvement of AI systems through high-quality data assessment and structured feedback.
