Role Responsibilities:
- Develop objective, verifiable evaluation criteria (rubrics) for system performance.
- Review system logs and execution paths to improve reliability and code quality.
- Refactor code and optimize system behavior toward ideal outcomes.
- Test systems for vulnerabilities, including data exposure and edge-case failures.
- Provide detailed, high-quality feedback on system performance and outputs.
Qualifications:
- 2+ years of experience in backend engineering, AI automation, or systems integration.
- Strong proficiency in at least two programming languages (e.g., Python, JavaScript, Go, Java).
- Experience working with SQL databases and building production-grade systems.
- Strong analytical skills and attention to detail for complex technical challenges.
- Experience in live, non-mocked environments with multi-step interactions.
Additional Information:
- Fully remote and flexible work schedule with project-based engagement and no guaranteed hours.
- Payment is based on completed tasks only, with freelancers able to accept or decline tasks based on availability.
- Workload may vary weekly, with no guaranteed volume and tasks assigned based on project availability.
Contract Company
The company is a contractor working with an enterprise client focused on training large language models for advanced AI systems. It is involved in a cutting-edge initiative to improve how AI writes, reviews, and optimizes code.