Work cross-functionally with Product and subject matter experts to conceptualize, prototype, and build data solutions
Connect disparate datasets (e.g. claims, contract rates, demographics data) to empower internal and external stakeholders
Build and maintain data engineering systems that support AI use cases, including scalable ingestion pipelines, feature generation, and downstream products
Turquoise Health aims to make healthcare pricing simpler, more transparent, and lower cost. They are a Series B startup backed by top VCs with an accomplished group of folks with a passion for improving healthcare.
Build and manage business data pipelines and transform Firefox telemetry data into structured datasets.
Partner with data scientists, product, and marketing teams to turn datasets into models and metrics.
Ensure data accuracy and performance using observability tools and resolve data issues.
Mozilla Corporation is a technology company backed by a non-profit that has shaped the internet, creating brands like Firefox. With millions of users globally, they focus on areas including AI and social media while remaining focused on making the internet better for people.
Design, implement, and maintain robust, scalable data pipelines to support AI, analytics, and operational reporting
Own and evolve the data warehouse architecture, ensuring it meets performance, flexibility, and governance needs
Ensure data integrity, availability, lineage, and observability across complex pipelines
Remote People is building the infrastructure to power borderless teams. Their technology handles global payroll, benefits, taxes, and compliance, enabling businesses to compliantly hire anyone anywhere at the push of a button. They are a growing, international family.
Architect, design, implement, and operate end-to-end data engineering solutions using Agile methodology.
Develop and manage robust data integrations with external vendors and organizations (including complex API integrations).
Collaborate closely with Data Analysts, Data Scientists, DBAs, and cross-functional teams to understand requirements and deliver high-impact data solutions.
SmartAsset is an online destination for consumer-focused financial information and advice, whose mission is helping people make smart financial decisions, reaching over an estimated 59 million people each month. A successful $110 million Series D funding round in 2021 valued the company at over $1 billion.
Design and develop high‑performance data converters for multi‑sensor autonomous‑driving data.
Design, build, and optimize large‑scale ingestion and transformation pipelines capable of processing petabyte‑scale autonomous‑driving sensor data.
Implement automated data validation, quality checks, and lineage tracking to ensure reliability of production datasets.
Torc has been a leader in autonomous driving since 2007 and is now part of the Daimler family. They are focused solely on developing software for automated trucks to transform how the world moves freight and have a collaborative, energetic, and team-focused culture.
Design, develop, and maintain scalable ETL/ELT pipelines for data ingestion.
Implement data quality checks, monitoring, and validation processes.
Automate manual processes into centralized and scalable solutions.
Informa TechTarget accelerates growth from R&D to ROI, informing and connecting technology buyers and sellers. They are a vibrant community of over 2000 colleagues worldwide and traded on Nasdaq as part of Informa PLC.
Design, build, and maintain scalable data pipelines.
Develop and optimize ETL/ELT processes using cloud data technologies.
Partner with teams to understand data requirements and improve data capture strategies.
Blueprint is a technology solutions firm with a strong presence across the United States, solving complicated problems for their clients. They are bold, smart, agile, and fun, and believe in unique perspectives, building teams of people with diverse skillsets and backgrounds.
Design, build, and maintain scalable data pipelines for clients across industries.
Architect and optimize cloud data warehouse solutions, adapting to each client's stack.
Collaborate with analysts and data scientists to ensure data is clean, reliable, and well-modeled.
NuView Analytics helps companies accelerate the time to insights from their data through data analytics, diligence, and fractional data science. They are a growth-stage company looking to drive additional value from the data they are sitting on and value humility, intellectual rigor, and stewardship.
Lead the architecture and evolution of scalable, distributed data pipelines, ensuring high availability and performance at scale
Build and maintain distributed web scraping systems using tools such as Playwright, Selenium, and BeautifulSoup
Integrate AI and LLMs into engineering workflows for code generation, automation, and optimization
MercatorAI is building scalable data infrastructure to power high-quality, data-driven decision making at scale. As an early-stage company, the team is focused on creating robust, future-ready systems that can handle complex data ingestion, transformation, and delivery across a growing national footprint.
Own organizational-wide data architecture, defining standards and designs.
Design and develop data pipelines, integrations, and platform features.
Partner with product managers to define new data features and capabilities.
They offer a connected equipment platform for managing mixed assets. The company values quality, continuous learning, and collaboration within a dynamic team environment.
Create innovative solutions for handling peta-bytes of data with billions of rows & joins.
Create real time and offline features generation pipelines to managing our data infrastructure to be reliable and fast!
Develop and productionize data pipelines for our ML models in both bare-metal and the cloud environment.
Kayzen is a mobile demand-side platform (DSP) dedicated to democratizing programmatic advertising. They enable leading apps, agencies, media buyers, and brands to run programmatic customer acquisition, retargeting, and brand performance campaigns through their self-serve and managed service options.
Manage and guide data teams to execute on enterprise data strategy.
Provide technical guidance and mentor team members on data technologies.
Design end-to-end data processing for enterprise data warehousing.
Jobgether uses an AI-powered matching process to ensure applications are reviewed quickly, objectively, and fairly against the role's core requirements. They identify the top-fitting candidates, and this shortlist is then shared directly with the hiring company.
Design and deploy scalable machine learning solutions to improve data quality, matching, classification, and automation.
Apply advanced analytics and generative AI to enhance data enrichment and ongoing data maintenance.
Drive analytical projects from problem definition through deployment and ongoing monitoring, in partnership with cross‑functional teams.
Experian is a global data and technology company, powering opportunities for people and businesses around the world. As a FTSE 100 Index company listed on the London Stock Exchange (EXPN), they have a team of 22,500 people across 32 countries and invest in people and new advanced technologies to unlock the power of data.
Lead the technical onboarding of partner institutions onto UDTS.
Design, build, and maintain scalable data pipelines and architectures.
Collaborate with team members to set engineering standards and guide data infrastructure strategy.
DataKind is a non-profit organization that uses data science and AI to address global challenges. They work with various sectors like health, humanitarian action, climate, economic opportunity, and education to create data-driven tools.
Help build scalable data solutions and streamline data ingestion.
Maintain high-quality databases that support our scientific and operational teams.
Optimize our data infrastructure to ensure efficient data access.
Funga is a public benefit corporation addressing the climate crisis by harnessing forest fungal networks. They are a team of passionate scientists and builders working to draw down at least three gigatons of carbon dioxide from the atmosphere by 2050.
Build and maintain robust data pipelines processing large volumes of data
Update and optimise our data platform for speed, scalability and cost
Develop processes and tools to monitor and analyse model performance and data accuracy
Moniepoint is Africa's all-in-one financial ecosystem, empowering businesses and their customers with seamless payment, banking, credit, and management tools. They processed $182 billion in 2023 and are Nigeria’s largest merchant acquirer, cultivating a culture of innovation, teamwork, and growth.
Implement the core components of our data platform, including data modeling, pipelines, and retrieval-ready storage layers.
Build operationally excellent systems, incorporating automated testing, deep observability, and robust failure handling.
Model complex business domains by defining clear entities, event histories, and reusable datasets that reflect real-world logic.
Campminder provides digital transformation solutions to the summer camp industry. The company has 85+ employees and is known for its values-led culture and employee experience, having been listed on Outside Magazine’s 50 Best Places to Work for 8 consecutive years.
Design and implement software systems that perform de-identification rules at high scale and throughput.
Generate and execute quality assurance plans to validate de-identification processes.
Run de-identification pipelines in health system cloud environments, optimizing for error rates and processing efficiency.
Dandelion Health, founded in 2020, is building the world’s largest AI training and clinical development platform, making data access easy for AI developers, pharma, and medical devices. Their culture emphasizes learning from data to help clients improve health through AI.
Collaborate with product managers, data analysts, and machine learning engineers to develop pipelines and ETL tasks.
Establish data architecture processes and practices that can be scheduled, automated, replicated and serve as standards.
Manage individual Data Engineers to foster learning, growth and success at Doximity.
Doximity is transforming the healthcare industry with a mission to help every physician be more productive and provide better care for their patients. As medicine's largest network in the United States, they are committed to building diverse teams with an inclusive culture.
Design, build, and maintain scalable data platforms using AWS to support analytics, machine learning, and emerging generative AI use cases.
Collaborate with data scientists, analysts, and engineering teams to translate business and AI requirements into scalable data solutions.
Work with large-scale datasets to build and optimize data pipelines using AWS services such as EMR (Spark, Trino), S3, Glue, Athena, and Airflow
Experian is a global data and technology company, powering opportunities for people and businesses around the world. They invest in people and new advanced technologies to unlock the power of data and to innovate. A FTSE 100 Index company listed on the London Stock Exchange, they have a team of 23,300 people across 32 countries.