AI Data Annotation & LLM Evaluation Projects (Outlier Platform)
Worked on multiple AI data annotation and evaluation projects through the Outlier platform, focusing on improving large language model performance and alignment. Tasks included reviewing and labeling text data, evaluating LLM-generated responses for correctness, relevance, and instruction-following, and classifying prompts based on defined guidelines. Ensured high-quality annotations by strictly adhering to project instructions, quality rubrics, and consistency standards. Maintained accuracy and attention to detail while handling complex and nuanced language tasks to support model training and evaluation at scale.