Similar Jobs

See all

What You'll Be Doing:

  • Own the design, implementation, and evolution of core MLOps systems across Hyperstack — including the infrastructure and workflows that underpin AI Studio
  • Build and improve systems that orchestrate model training, fine-tuning, evaluation, and deployment — engineered for long-running, resource-intensive, GPU workloads
  • Own production readiness across ML infrastructure — monitoring, alerting, incident response, and continuous improvement based on real-world usage

About You:

  • Proven experience designing, building, and operating production ML infrastructure, platform systems, or MLOps workflows in cloud environments
  • Hands-on Python development skills, with experience building backend systems, automation, and developer or platform tooling
  • Experience supporting LLM, generative AI, or fine-tuning workflows in production — including training, evaluation, deployment, inference, and lifecycle management

What We Offer:

  • Competitive salary and annual discretionary bonus scheme
  • 25 days of holiday, plus public holidays
  • Flexible working arrangements (remote or hybrid, depending on role and location)

NexGen Cloud

NexGen Cloud is the company behind Hyperstack, a full-stack AI cloud serving tens of thousands of customers from AI researchers to enterprises running the world's most compute-intensive workloads. They deliver on-demand and private GPU infrastructure to teams who treat performance as a requirement, not a feature.

Apply for This Position