Design, build, and maintain scalable data platforms using AWS to support analytics, machine learning, and emerging generative AI use cases.
Collaborate with data scientists, analysts, and engineering teams to translate business and AI requirements into scalable data solutions.
Work with large-scale datasets to build and optimize data pipelines using AWS services such as EMR (Spark, Trino), S3, Glue, Athena, and Airflow
Experian is a global data and technology company, powering opportunities for people and businesses around the world. They invest in people and new advanced technologies to unlock the power of data and to innovate. A FTSE 100 Index company listed on the London Stock Exchange, they have a team of 23,300 people across 32 countries.
Design Scalable Data Architecture: Build modern, cloud-native data platforms (AWS, Snowflake, Databricks) supporting batch and streaming use cases.
Develop Efficient Data Pipelines & Models: Automate ETL/ELT workflows, optimise data models, and enable self-serve analytics and AI.
End-to-End Data Ownership: Manage ingestion, storage, processing, and delivery of structured and unstructured data.
Trustonic makes smartphones affordable, enabling global access to devices and digital finance through secure smartphone locking technology. They partner with mobile carriers, retailers, and financiers across 30+ countries, powering device financing solutions. The company celebrates its diversity and is looking to do the right thing: for each other, the community and the planet.
Build and lead a team of 4-5 data engineers focused on reusable product artifacts
Own the product data engineering backlog in partnership with product management
Define and enforce technical standards for notebooks, pipelines, QC modules, and documentation
Qualified Health is redefining what’s possible with Generative AI in healthcare. They provide the guardrails for safe AI governance, healthcare-specific agent creation, and real-time algorithm monitoring, working alongside leading health systems to drive real change. They are a fast-growing company backed by premier investors.
Own and evolve the data infrastructure that powers Clever's core data products.
Maintain and improve data pipeline reliability, monitoring and resolving pipeline failures.
Design and implement ingestion for new operational data sources that support Clever's speed-to-match initiative.
Clever Real Estate is a venture-backed technology company aiming to revolutionize real estate transactions. They have built a leading online education platform helping consumers save money and have earned a 4.9 TrustPilot rating with over 3,800 reviews.
Architect and sustain self-healing pipelines using Astronomer/Airflow to ensure 24/7 data availability.
Design and optimize event-driven API ingestion frameworks leveraging AWS Lambda and DLT (Data Load Tool).
Manage high-performance modeling within AWS Redshift, utilizing DBT to transform raw transactional data into high-fidelity business intelligence.
Odisea helps close the opportunity gap between Colombia and the United States by redefining nearshoring. They are building a passionate team of professionals committed to this purpose.
Collaborate with data scientists, analysts, and other stakeholders to understand data requirements and design data models and schemas that facilitate data analysis and reporting
Design, develop, and maintain scalable and efficient data pipelines and ETL processes to ingest, process, and transform large volumes of data from various sources into usable formats
Build and optimize data storage and processing systems, including data warehouses, data lakes, and big data platforms, using AWS services such as Amazon Redshift, AWS Glue, AWS EMR, AWS S3, and AWS Lambda, to enable efficient data retrieval and analysis
ATPCO is the world's primary source for air fare content, holding over 200 million fares across 160 countries. Every day, the travel industry relies on ATPCO's technology and data solutions to help millions of travelers reach their destinations efficiently. At ATPCO, they believe in flexibility, trust, and a culture where your wellbeing comes first.
Lead and grow a team of data engineers, providing mentorship and technical guidance.
Own execution of customer integrations across multiple product lines, ensuring on-time delivery.
Improve data quality and pipeline reliability by investing in better alerting and resilience.
Afresh is the leading AI company in fresh food, partnering with grocers to order billions of dollars of fresh food. They are on a mission to eliminate food waste and make fresh food accessible to all and has saved 200M lbs of food waste in 2025 alone.
Serve as the embedded technical lead for Databricks customer engagements.
Own Databricks platform architecture, design decisions, and technical standards.
Lead delivery of complex data pipelines and analytics workloads on Databricks.
540 is a forward-thinking company that the government turns to in order to #getshitdone. They break down barriers, build impactful technology, and solve mission-critical problems.
Manage and support a team of 6 Data Engineers, helping them focus on impactful technical initiatives as the platform scales.
Drive execution excellence, ensuring the team delivers with high velocity, quality, and reliability.
Foster a healthy and sustainable team environment by helping the team manage workload and focus on meaningful engineering work.
Bluefish believes that AI represents the next major chapter of the internet and that consumers will increasingly use AI to consume information and media online. Bluefish is building the platform that helps brands engage consumers on this new AI channel, with powerful enterprise tools to manage AI brand safety and engage consumers with thoughtful and personalized AI marketing experiences.
Lead the implementation of a resilient, privacy-first data platform architecture.
Lead the design, infrastructure, and tooling decisions for platform optimization.
Develop AI-ready architecture by creating semantic layers that define and standardize business logic.
Headspace provides access to lifelong mental health support. They combine evidence-based content, clinical care, and innovative technology to help millions of members around the world get support that’s effective and personalized. They value connecting with courage, ownership, and iterating to great.
Design and implement scalable data architectures to support business needs.
Build and optimize data pipelines, ensuring data accessibility and security.
Develop and maintain data models, databases, and data lakes, with robust data governance.
Terawatt Infrastructure delivers large scale, turnkey charging solutions for companies rapidly deploying AV and EV fleets. With a growing portfolio of sites across the US, Terawatt is building the permanent transportation and logistics infrastructure of tomorrow through capital, real estate, development, and site operations solutions.
Design and implement scalable data ingestion and transformation pipelines using Databricks and cloud platforms
Lead architecture decisions for modern data platforms, including Medallion Architecture and Lakehouse patterns
Build and maintain ETL/ELT pipelines using Python and SQL, following engineering best practices
AOT Technologies helps enterprises and governments bring their ideas to life. As a boutique consulting firm, they partner with enterprises, startups, and governments to solve complex, mission-critical challenges. Their teams are collaborative and their leadership is transparent.
Architect, build, and refine our infrastructure and tools that enable other teams to work with data.
Efficiently handle vast amounts of clickstream, database, and third party application data, processing in batch and real-time
Take full ownership of the solutions you build, working with stakeholders to develop requirements, implement solutions, monitor production, and troubleshoot problems that arise.
Gusto is on a mission to grow the small business economy. They handle payroll, health insurance, 401(k)s, and HR, so owners can focus on their craft and customers. Gusto supports more than 400,000 small businesses across the country, and they’re building a workplace that represents and celebrates the customers they serve.
Design, build, and maintain scalable batch and real-time data pipelines that power analytics, experimentation, and machine learning
Partner cross-functionally with analytics, product, engineering and operations to deliver high-quality data solutions that drive measurable business impact
Champion data quality, reliability, and observability by implementing best practices in testing, monitoring, lineage, and incident response
Gopuff is reimagining how people purchase everyday essentials, from snacks to household goods to alcohol, all delivered in minutes. They are assembling a team of thinkers, dreamers and risk-takers who know the value of peace of mind in an unpredictable world.
Create innovative solutions for handling peta-bytes of data with billions of rows & joins.
Create real time and offline features generation pipelines to managing our data infrastructure to be reliable and fast!
Develop and productionize data pipelines for our ML models in both bare-metal and the cloud environment.
Kayzen is a mobile demand-side platform (DSP) dedicated to democratizing programmatic advertising. They enable leading apps, agencies, media buyers, and brands to run programmatic customer acquisition, retargeting, and brand performance campaigns through their self-serve and managed service options.
Design, implement, and maintain robust, scalable data pipelines to support AI, analytics, and operational reporting
Own and evolve the data warehouse architecture, ensuring it meets performance, flexibility, and governance needs
Ensure data integrity, availability, lineage, and observability across complex pipelines
Remote People is building the infrastructure to power borderless teams. Their technology handles global payroll, benefits, taxes, and compliance, enabling businesses to compliantly hire anyone anywhere at the push of a button. They are a growing, international family.
Design, develop, and maintain scalable ETL/ELT pipelines for data ingestion.
Implement data quality checks, monitoring, and validation processes.
Automate manual processes into centralized and scalable solutions.
Informa TechTarget accelerates growth from R&D to ROI, informing and connecting technology buyers and sellers. They are a vibrant community of over 2000 colleagues worldwide and traded on Nasdaq as part of Informa PLC.
Architect, implement, and maintain scalable data architectures.
Develop, optimize, and maintain ETL processes.
Optimize data processing and query performance.
Blueprint Technologies is a technology solutions firm that helps organizations unlock value from existing assets by leveraging cutting-edge technology. Their teams have unique perspectives and years of experience across multiple industries. They believe in unique perspectives and build teams of people with diverse skillsets and backgrounds.
Design and evolve scalable data pipelines and architectures.
Act as the primary anchor for data ingestion, transformation, and storage solutions.
Ensure mission-critical data is accessible and reliable.
CodeRoad provides end-to-end software development services, helping businesses scale with ideal infrastructure solutions. From staff augmentation to dedicated IT teams and general software engineering, their nearshore technology services empower businesses to thrive in an ever-evolving digital landscape.
Utilize SQL and data transformation tools (like DBT and Coalesce) to develop and deploy abstracted data models following the “Medallion” concept.
Collaborate with other engineering teams as well as business analysts and end-users to develop requirements and design useful data models.
Develop DAGs in Apache Airflow for delivering data products to internal and external partners.
Forbright is a nationwide full-service bank and commercial lender focused on helping to build a brighter future for our clients and the communities we serve. They are a dynamic, high energy, fun, and fast-paced organization that has an exciting growth trajectory.