Design and implement tooling that enables researchers to quickly deploy and evaluate new models in production
Design, build, and maintain high-performance, cost-efficient inference pipelines, making architectural decisions about scaling, reliability, and cost trade-offs
Proactively identify and resolve infrastructure bottlenecks, proposing and scoping improvements to iteration speed and production reliability
AssemblyAI builds best-in-class Speech AI models that power the next generation of voice applications. They are a remote team building one of the next great AI companies where teammates define and build their company culture.
Contribute to the development of the Everywhere Inference platform, a Kubernetes-based solution.
Design and implement APIs and developer tools to simplify deployment, management, and monitoring of AI applications.
Focus on packaging and integrating new ML models into the platform, using Python and common ML frameworks.
Gcore provides infrastructure and software solutions for AI, cloud, network, and security. They power everything from real-time communication and streaming to enterprise AI and secure web applications, with over 550 professionals globally and partnerships with technology leaders.
Contribute to the development of the Everywhere Inference platform, a Kubernetes-based solution.
Design and implement APIs and developer tools to simplify deployment, management, and monitoring of AI applications.
Optimize serverless container workflows for AI workloads, ensuring performance, scalability, and seamless autoscaling.
Gcore provides infrastructure and software solutions for AI, cloud, network, and security. They have 550+ professionals globally and collaborate with technology partners such as Intel, NVIDIA, Dell, and Equinix.
Design, build, and manage our cloud infrastructure using modern tools (Pulumi) to ensure all infrastructure changes are reproducible, secure, and easily auditable.
Orchestrate and optimize our Kubernetes clusters for complex, compute-heavy AI workloads, guaranteeing maximum efficiency and fault tolerance.
Implement a flawless monitoring setup using Datadog and OpenTelemetry to make the black box of our distributed systems transparent, hunting down latency spikes or bottlenecks before they impact users.
Deepslate is building Speech to Speech Voice AI models that sound and act indistinguishable from a human, with the belief that everyone should be able to use it. Backed by top-tier investors from the Tech and AI sectors, we are incredibly well-funded and moving fast.
Architect and optimize distributed training and inference systems for large-scale AI models
Design and deliver customer-focused solutions that maximize performance and business value
Lead the transition of ML pipelines from POC to scalable production systems
The company offers an AI-centric cloud platform reshaping the landscape of artificial intelligence. They provide infrastructure, tools, and services for developers to service the explosive growth of the global AI industry, catering to Fortune 1000 companies, startups, and AI researchers.
Set up and manage GPU cluster infrastructure on major cloud providers.
Build and operate job orchestration and scheduling systems.
Integrate and maintain ML training frameworks and post-training pipelines.
Snorkel AI helps enterprises transform expert knowledge into specialized AI at scale. They started as a research project in the Stanford AI Lab and work with some of the world’s largest organizations to empower scientists, engineers, financial experts, product creators, journalists, and more to build custom AI with their data faster than ever before.
Own model serving: Design, build, and maintain low-latency, highly-available serving stacks for in-house ML model serving and integrating with LLM serving partners.
Automate training pipelines: Orchestrate data prep, training, evaluation, and registry workflows on Kubernetes with solid MLOps practices.
Optimize at scale: Profile and tune throughput, memory, and cost; introduce caching, sharding, batching, and GPU/CPU autoscaling where it pays off.
Cresta aims to turn every customer conversation into a competitive advantage by unlocking the true potential of the contact center. Their platform combines AI and human intelligence to help contact centers discover customer insights and automate conversations.
Standardize CI/CD pipelines (GitHub Actions) and Helm charts across 10+ microservices
Build centralized logging, metrics, and alerting (currently a gap)
Extend Terraform to cover full AWS infrastructure
Kiefer Tech delivers cutting-edge AI, robotics, and enterprise solutions across Greece and the EU, leveraging over 20 years of engineering heritage from the Green Energy sector. As the technology arm of Kiefer, they are guided by innovation, quality, and long-term client partnerships and are building sovereign AI infrastructure.
Design and deploy high-performance agentic systems that leverage Fastino’s optimized model architectures.
Collaborate with engineering teams to turn novel architectural breakthroughs into scalable solutions for enterprise customers.
Drive rapid, iterative prototyping of AI functionalities, refining model performance and task-accuracy based on real-world telemetry.
Fastino is building the next generation of LLMs with a team of alumni from Google Research, Apple, Stanford, and Cambridge and has developed the GLiNER family of open source models. Fastino has raised $25M through seed round and is backed by leading investors including Microsoft, Khosla Ventures, and Insight Partners.
Provide and own automation of the provisioning of CSP resources, including networking, Kubernetes clusters and specific CSP resources required by our application teams.
Work with users (Grafana Cloud application teams) to help understand their needs and ensure investment in the right capabilities.
Participate in the Platform department Infrastructure wing on-call rotation.
Grafana Labs is a remote-first, open-source powerhouse with more than 20M users of Grafana around the globe. The team thrives in an innovation-driven environment where transparency, autonomy, and trust fuel everything that they do.
Design and implement production-scale AI agent systems and orchestration frameworks.
Deploy and optimize LLMs/SLMs in production with fine-tuning techniques.
Build data pipelines for training data curation, synthetic generation, and PII masking.
FourKites is the leader in AI-driven supply chain transformation for global enterprises and a pioneer of advanced real-time visibility. They turn supply chain data into automated action helping 1,600+ global brands prevent disruptions. They provide competitive compensation with stock options, outstanding benefits and a collaborative culture for all employees around the globe.
Design and build the infrastructure layer powering AI agent systems in production
Develop high-performance Rust services that handle model inference, orchestration, and execution
Architect scalable systems capable of supporting millions of users and high request throughput
Kraken is a mission-focused company rooted in crypto values, aiming to accelerate the global adoption of crypto so that everyone can achieve financial freedom and inclusion. As a fully remote company, Kraken has employees in 70+ countries and is committed to industry-leading security, crypto education, and client support.
Build, operate, and continuously evolve ultra-resilient, cloud-native platforms utilizing Kubernetes, Docker, and advanced container orchestration.
Eliminate manual toil by engineering heavily automated infrastructure using robust Infrastructure as Code (IaC) tools like Terraform and Ansible.
Implement and operate cutting-edge CI/CD pipelines optimized for the rapid, secure deployment of mission-critical software, APIs, and AI/ML models.
Deutsche Telekom IT Solutions is a subsidiary of the Deutsche Telekom Group and was Hungary’s most attractive employer in 2025, according to Randstad’s representative survey. The company provides a wide portfolio of IT and telecommunications services with more than 5300 employees.
Build and manage the full ML lifecycle—from experiment tracking to model deployment and retraining.
Implement ML-specific CI/CD (e.g., CML, Kubeflow Pipelines) to automate the promotion of models to production.
Architect distributed systems for large-scale model inference.
Deutsche Telekom IT Solutions is a subsidiary of the Deutsche Telekom Group, recognized as Hungary’s most attractive employer in 2025. They provide IT and telecommunications services with more than 5300 employees, serving hundreds of large customers in Germany and other European countries.
Design, develop, and maintain high-quality software solutions using Python.
Contribute to the design and evolution of scalable and maintainable software architectures.
Deploy, operate, and monitor applications in cloud environments (AWS, Azure, or GCP).
Lynx Analytics works on real-world AI and advanced analytics solutions with measurable business impact. They have a collaborative culture that values real outcomes, offering high ownership and rapid learning opportunities.
Own the container-based application lifecycle, bi-weekly releases, and CI/CD pipelines for GMS.
Manage deployments on customer-isolated Kubernetes clusters running stateful applications.
Ensure high availability and performance by meeting contractual SLAs through proactive monitoring and alert response.
Planet designs, builds, and operates the largest constellation of imaging satellites in history, delivering data via a cloud-based platform. They are both a space company and data company with a people-centric approach, striving to put their team members first.
Architect and scale the core intelligence behind our platform.
Design, build, and optimize the pipelines and agent systems that drive live customer interactions.
Build real-time and batch pipelines for ingestion, training, and inference.
Raynmaker is building RaynBrain, an agentic AI platform for complex conversations grounded in machine learning, neuroscience, and forensic linguistics. They empower autonomous systems that interpret, adapt, and act in real time, turning raw leads into revenue without scripts or human handoffs. Raynmaker is a small team helping other small teams move faster and convert more leads.
Secure AI-specific attack surfaces: prompt injection defenses, PII handling in LLM pipelines, model interaction data leakage
Kiefer Tech leverages over 20 years of engineering heritage from the Green Energy sector to deliver cutting-edge AI, robotics, and enterprise solutions across Greece and the EU. They build sovereign AI infrastructure that keeps data within EU borders, respect privacy, and delivers tangible business impact.
Design and build AI solutions for real-world business problems.
Own the end-to-end lifecycle of AI features.
Develop and optimize LLM-based applications.
Remote People is building the infrastructure to power borderless teams. Their technology handles global payroll, benefits, taxes, and compliance, enabling businesses to hire anyone anywhere compliantly at the push of a button. They are committed to building a global, diverse team representing different and varied backgrounds, perspectives, and experiences.
Serve as Zencore’s senior-most technical authority on the practical application of advanced artificial intelligence and machine learning.
Partner with the sales and business development teams in a pre-sales capacity to scope opportunities, design solutions for proposals, and act as the senior technical voice in client pitches.
Lead the architecture and design of sophisticated, secure, and scalable AI solutions for our clients, moving beyond standard API integrations to create genuine competitive advantages.
Zencore is a fast-growing company founded by former Google Cloud leaders, architects, and engineers. Our engagements eliminate obstacles, reduce risk, and accelerate timelines for customers adopting and scaling modern AI solutions.