LLM Text Evaluation and Prompt Optimization Project
Participated in evaluating and refining outputs from large language models (LLMs) used for text generation and summarization tasks. Annotated datasets with accuracy and tone ratings, corrected factual inconsistencies, and created high-quality prompt–response pairs for supervised fine-tuning. Used AWS SageMaker and Label Studio for data management, ensuring consistent formatting, labeling accuracy, and adherence to quality assurance guidelines. The project involved thousands of text samples and contributed to improved model coherence and factual reliability.