Skip to main content

LLM Based Automated Code Synthesis & Evaluation Pipeline

Python, Java, pytest, Large Language Models (LLM), JSON, Pandas, SciPy, Shell Scripting, Git, Prompt Engineering

  • Developed an end-to-end automation pipeline using LLM's for generating programming problem datasets and selecting a reproducible subset for evaluation.
  • Engineered systems to interact with large language models for both code synthesis and translation tasks, including experiments with advanced quantization configurations.
  • Implemented automated prompting for detecting bugs and generating detailed code coverage reports to enhance code quality.
  • Designed integrated validation routines to verify the integrity of generated datasets, model outputs, and essential project artifacts.
  • Built command-line scripts and automated workflows to seamlessly orchestrate dataset preparation, model prompting, and result evaluation, ensuring reproducibility and reliability.
  • Contributed to the project's impact by streamlining the process of code evaluation and translation, thus enabling robust experimental research and development in automated code synthesis.