Extend, optimize, and maintain core data models for reports, machine learning, and generative AI.
Implement automation and operationalize ML models to streamline operational processes and improve efficiency.
Partner with engineering, product, and analytics teams to deliver seamless integrations and customer-facing data products.
Boulevard provides a client experience platform for appointment-based, self-care businesses, helping customers enhance client experiences. They value diversity and inclusivity, offering equal opportunities and aiming to create a supportive work environment.
Build and manage business data pipelines and transform Firefox telemetry data into structured datasets.
Partner with data scientists, product, and marketing teams to turn datasets into models and metrics.
Ensure data accuracy and performance using observability tools and resolve data issues.
Mozilla Corporation is a technology company backed by a non-profit that has shaped the internet, creating brands like Firefox. With millions of users globally, they focus on areas including AI and social media while remaining focused on making the internet better for people.
Help customers love Rerun, being the primary technical interface.
Run demos, build engineer-to-engineer trust, and support onboarding.
Map customer needs to product roadmap and build features based on understanding.
Rerun is building the data stack for Physical AI. They have an uncommonly talented tech team and expect everyone to take broad responsibility for what they build.
Build and maintain robust data pipelines processing large volumes of data
Update and optimise our data platform for speed, scalability and cost
Develop processes and tools to monitor and analyse model performance and data accuracy
Moniepoint is Africa's all-in-one financial ecosystem, empowering businesses and their customers with seamless payment, banking, credit, and management tools. They processed $182 billion in 2023 and are Nigeria’s largest merchant acquirer, cultivating a culture of innovation, teamwork, and growth.
Work cross-functionally with Product and subject matter experts to conceptualize, prototype, and build data solutions
Connect disparate datasets (e.g. claims, contract rates, demographics data) to empower internal and external stakeholders
Build and maintain data engineering systems that support AI use cases, including scalable ingestion pipelines, feature generation, and downstream products
Turquoise Health aims to make healthcare pricing simpler, more transparent, and lower cost. They are a Series B startup backed by top VCs with an accomplished group of folks with a passion for improving healthcare.
Design and build the AI platform layer, including data pipelines and serving infrastructure.
Productionize AI/ML capabilities, ensuring reliability, performance, and scalability.
Architect data pipelines to ingest, transform, and serve data to power AI features.
Lone Wolf Technologies is building AI capabilities into the core of its platform, transforming how real estate professionals manage transactions, serve clients, and grow their businesses. The company's culture fosters innovation and collaboration, empowering employees to contribute to impactful projects.
Build a high performing team by hiring and nurturing engineering talent.
Drive technical solutioning and building roadmaps.
Work closely with engineering leaders to drive engineering excellence in our processes and systems.
Aledade empowers primary care physicians with technology to keep their patients healthy, preventing unnecessary hospitalizations. They are a technology company that helps primary care doctors deliver better care at a lower cost.
Conduct end-to-end analyses, wrangling data via SQL or Python, to statistical modeling, to hypothesizing and presenting business ideas.
Mentor and guide junior engineers, fostering a culture of learning and technical excellence.
Support the development and deployment of projects involving machine learned models for offline, batch-based data products as well as models deployed to online, real-time services.
Yelp's mission is connecting people with great local businesses, requiring the use of cutting-edge Machine Learning (ML) and Artificial Intelligence (AI). Their engineering culture values individual authenticity and encourages creative solutions, supporting growth in a collaborative environment.
Design, implement, and maintain robust, scalable data pipelines to support AI, analytics, and operational reporting
Own and evolve the data warehouse architecture, ensuring it meets performance, flexibility, and governance needs
Ensure data integrity, availability, lineage, and observability across complex pipelines
Remote People is building the infrastructure to power borderless teams. Their technology handles global payroll, benefits, taxes, and compliance, enabling businesses to compliantly hire anyone anywhere at the push of a button. They are a growing, international family.
Lead and develop a high-performing Data Science and Analytics Engineering team
Drive data-led product improvements across growth, engagement, and monetization
Partner closely with Product, Engineering, and Leadership to shape strategy
ClassDojo aims to provide every child globally with an education they love, starting with a communication network. They connect K-12 teachers, children, and families, reaching over 45 million children in 180 countries with a team of around 200, fostering a talented, entrepreneurial, and innovative culture.
Collaborate with product managers, data analysts, and machine learning engineers to develop pipelines and ETL tasks.
Establish data architecture processes and practices that can be scheduled, automated, replicated and serve as standards.
Manage individual Data Engineers to foster learning, growth and success at Doximity.
Doximity is transforming the healthcare industry with a mission to help every physician be more productive and provide better care for their patients. As medicine's largest network in the United States, they are committed to building diverse teams with an inclusive culture.
Implement the core components of our data platform, including data modeling, pipelines, and retrieval-ready storage layers.
Build operationally excellent systems, incorporating automated testing, deep observability, and robust failure handling.
Model complex business domains by defining clear entities, event histories, and reusable datasets that reflect real-world logic.
Campminder provides digital transformation solutions to the summer camp industry. The company has 85+ employees and is known for its values-led culture and employee experience, having been listed on Outside Magazine’s 50 Best Places to Work for 8 consecutive years.
You have 7+ years of experience in Data Science and a track record of leading high-complexity projects from scoping to production with minimal direction.
You can deploy and iterate ML and NLP solutions in production environments, with working knowledge of LLMs including fine-tuning, LLM-as-judge, and AI-assisted development.
You write production-grade code in Python or a comparable high-level language.
DuckDuckGo is an online protection company with a remote-first team of 300+ employees. Their mission is to raise the standard of trust online, and they have a culture of trust, inclusivity, and empowered project management.
Contribute to applied machine learning and data initiatives that directly support product capabilities.
Assist in building or improving data pipelines, feature engineering processes, or model evaluation frameworks.
Participate in prototyping AI-enabled product features such as recommendations, automation workflows, or predictive insights.
Pacvue helps brands and agencies succeed in retail media and commerce through intelligent systems. They are committed to employing a diverse workforce, valuing qualified applicants regardless of various backgrounds and identities.
Design, build, and maintain data products that support R&D, analytics, Lab, and scientific workflows.
Build and maintain data pipelines for large and complex datasets ensuring high data quality.
Partner with scientists and engineers to translate research needs into reusable data assets.
Natera is a global leader in cell-free DNA (cfDNA) testing, dedicated to oncology, women’s health, and organ health. They aim to make personalized genetic testing and diagnostics part of the standard of care to protect health and enable earlier and more targeted interventions that lead to longer, healthier lives.
Creatively writing prompts and responses to a variety of diverse topics.
Leading labeling initiatives with third party firms and internal customers.
Creating and updating detailed guidelines and specifications for stakeholders.
Welo Data provides AI services, specifically data annotation. They enable brands and companies to reach, engage, and grow international audiences, delivering multilingual content transformation services in translation, localization, and adaptation.
Design and develop an AI-powered productivity analytics platform.
Build scalable LLM pipelines and create a meta-workflow system.
Develop system-level prompt engineering and build an evaluation framework for AI output quality control.
Appflame is a Ukrainian product-driven tech company committed to building world-class products. They have 500+ team members and offices in Kyiv, London, Limassol, and a co-working hub in Warsaw; they value bold, driven people who are passionate about building real products.
Lead and grow a team of data engineers, providing mentorship and technical guidance.
Own execution of customer integrations across multiple product lines, ensuring on-time delivery.
Improve data quality and pipeline reliability by investing in better alerting and resilience.
Afresh is the leading AI company in fresh food, partnering with grocers to order billions of dollars of fresh food. They are on a mission to eliminate food waste and make fresh food accessible to all and has saved 200M lbs of food waste in 2025 alone.
Lead the architecture and evolution of scalable, distributed data pipelines, ensuring high availability and performance at scale
Build and maintain distributed web scraping systems using tools such as Playwright, Selenium, and BeautifulSoup
Integrate AI and LLMs into engineering workflows for code generation, automation, and optimization
MercatorAI is building scalable data infrastructure to power high-quality, data-driven decision making at scale. As an early-stage company, the team is focused on creating robust, future-ready systems that can handle complex data ingestion, transformation, and delivery across a growing national footprint.