Role Responsibilities:

  • Develop objective, verifiable evaluation criteria (rubrics) for system performance.
  • Review system logs and execution paths to improve reliability and code quality.
  • Refactor code and optimize system behavior toward ideal outcomes.
  • Test systems for vulnerabilities, including data exposure and edge-case failures.
  • Provide detailed, high-quality feedback on system performance and outputs.

Qualifications:

  • 2+ years of experience in backend engineering, AI automation, or systems integration.
  • Strong proficiency in at least two programming languages (e.g., Python, JavaScript, Go, Java).
  • Experience working with SQL databases and building production-grade systems.
  • Strong analytical skills and attention to detail for complex technical challenges.
  • Experience in live, non-mocked environments with multi-step interactions.

Additional Information:

  • Fully remote and flexible work schedule with project-based engagement and no guaranteed hours.
  • Payment is based on completed tasks only, with freelancers able to accept or decline tasks based on availability.
  • Workload may vary weekly, with no guaranteed volume and tasks assigned based on project availability.

Contract Company

The company is a contractor working with an enterprise client focused on training large language models for advanced AI systems. It is involved in a cutting-edge initiative to improve how AI writes, reviews, and optimizes code.

Apply for This Position