Build & Operate Data Pipelines, using AWS-native data tools and distributed processing frameworks.
Operate and improve core data platform services, addressing incidents, performance issues, and operational toil.
Partner with data producers and consumers to onboard pipelines, troubleshoot issues, and improve platform usability.
Fetch is a platform where millions of people use Fetch earning rewards for buying brands they love, and a whole lot more. With investments from SoftBank, Univision, and Hamilton Lane, and partnerships with Fortune 500 companies, it is reshaping how brands and consumers connect in the marketplace. Ranked as one of America’s Best Startup Employers by Forbes, Fetch fosters a people-first culture rooted in trust, accountability, and innovation.
Design, build, and maintain a robust, self-service, scalable, and secure data platform. Create and edit data pipelines, considering business logic, levels of aggregation, and data quality. Enable teams to access and use data effectively through self-service tools and well-modeled datasets.
We are Grupo QuintoAndar, the largest real estate ecosystem in Latin America, with a diversified portfolio of brands and solutions across different countries.
Strengthen the real estate MLS data platform squad. Build robust data pipelines and backend services. Own the end-to-end architecture for MLS and property data.
Luxury Presence is the leading digital platform revolutionizing the real estate industry for agents, teams, and brokerages.
Architect and maintain robust data pipelines to transform diverse data inputs.
Integrate data from various sources into a unified platform.
Build APIs with AI assistance to enable secure access to consolidated insights.
Abusix is committed to making the internet a safer place. They are a globally distributed team that spans multiple countries and thrives in a culture rooted in trust, ownership, and collaboration.
Enable efficient consumption of domain data as a product by delivering and promoting strategically designed actionable datasets and data models
Build, maintain, and improve rock-solid data pipelines using a broad range of technologies like AWS Redshift, Trino, Spark, Airflow, and Kafka streaming for real-time processing
Support teams without data engineers in building decentralised data solutions and product integrations, for example, around DynamoDB Act as a data ambassador, promoting the value of data and our data platform among engineering teams and enabling cooperation
OLX operates consumer brands that facilitate trade to build a more sustainable world. They have colleagues around the world who serve millions of people every month.
Partner with our customer teams to develop engineering plans to implement our health system partners
Build and support robust batch and streaming pipelines
Evolve the maturity of our monitoring systems and processes to improve visibility and failure detection in our infrastructure
Paradigm is rebuilding the clinical research ecosystem by enabling equitable access to trials for all patients. Incubated by ARCH Venture Partners and backed by leading healthcare and life sciences investors, Paradigm’s seamless infrastructure implemented at healthcare provider organizations, will bring potentially life-saving therapies to patients faster.
Design, build, and scale performant data pipelines and infrastructure, primarily using ClickHouse, Python, and dbt.
Build systems that handle large-scale streaming and batch data, with a strong emphasis on correctness and operational stability.
Own the end-to-end lifecycle of data pipelines, from raw ingestion to clean, well-defined datasets consumed by downstream teams.
Nansen is a leading blockchain analytics platform that empowers investors and professionals with real-time, actionable insights derived from on-chain data. We’re building the world’s best blockchain analytics platform, and data is at the heart of everything we do.
Collaborate with engineering, data science, ML, data engineering, and product analytics teams to understand and shape the future needs of our data platform and infrastructure.
Define, drive, and implement the future live ingestion layer of data into our data platform (e.g. Kafka, Kinesis).
Define and evolve standards for storage, compute, data management, provenance, and orchestration.
Inspiren offers the most complete and connected ecosystem in senior living.
Lead product requirements and advanced analytics requirements gathering efforts.
Work with analytics, data science, and wider engineering teams to help with automating data analysis and visualization needs.
Build a scalable technology platform to support a growing business and deliver high-quality code to production.
Achieve is a leading digital personal finance company that helps everyday people move from struggling to thriving by providing innovative, personalized financial solutions. They have over 3,000 employees in mostly hybrid and 100% remote roles across the United States with hubs in Arizona, California, and Texas and a culture of putting people first.
Design and build scalable data pipelines that ingest, process, and transform high-volume event streams and historical data
Develop and maintain APIs that deliver analytics, trend reports, and drill-down capabilities to internal teams and external customers
Build robust infrastructure for data quality monitoring, ensuring accuracy and completeness across customer and artifact datasets
Socket helps devs and security teams ship faster by cutting out security busywork. They have raised $65M in funding from top angels, operators, and security leaders.
Architect and lead the evolution of our modern data platform.
Design and build production LLM pipelines and infrastructure that power intelligent operations.
Own end-to-end data acquisition and integration architecture across diverse sources.
Brightwheel is the largest, fastest growing, and most loved platform in early ed. They are trusted by millions of educators and families every day. The team is passionate, talented, and customer-focused and embodies their Leadership Principles in their work and culture.
Design and develop scalable data pipelines and infrastructure to process large volumes of data efficiently
Collaborate with cross-functional teams to ensure data integrity, accessibility, and usability
Implement and maintain data quality measures throughout the data lifecycle
CI&T is a tech transformation specialist, uniting human expertise with AI to create scalable tech solutions. With over 8,000 employees around the world, they have a culture that values diverse identities and life experiences, fostering a diverse, inclusive, and safe work environment.
Design and develop distributed systems, participating in architectural decisions.
Assume ownership of components within our blockchain data platform, such as data ingestion.
Strive for automating processes by leveraging AI and building agents.
Dune is on a mission to make crypto data accessible through their collaborative multi-chain analytics platform. They are a team of approximately 60 employees working together across Europe and the eastern US, backed by world-class investors, aiming to empower onchain analysts.
Support production systems and help triage issues during live sporting events
Architect low-latency, real-time analytics systems including raw data collection, feature development and endpoint production
Integrate large and complex real-time datasets into new consumer and enterprise products
Swish Analytics is a sports analytics, betting, and fantasy startup building the next generation of predictive sports analytics data products. They're looking for team-oriented individuals with an authentic passion for accurate and predictive real-time data who can execute in a fast-paced, creative, and continually-evolving environment without sacrificing technical excellence.
Design, build, and maintain highly scalable, reliable, and efficient ETL/ELT pipelines.
Ingest data from a multitude of sources and transform raw data into clean, structured, and AI/ML-ready formats.
Work closely with data scientists, machine learning engineers, and business analysts to understand their data needs.
Valtech exists to unlock a better way to experience the world by blending crafts, categories, and cultures, helping brands unlock new value in an increasingly digital world.
Design, implement, and maintain distributed ingestion pipelines for structured and unstructured data.
Build scalable ETL/ELT workflows to transform, validate, and enrich datasets for AI/ML model training and analytics.
Support preprocessing of unstructured assets for training pipelines, including format conversion, normalization, augmentation, and metadata extraction.
Meshy is a leading 3D generative AI company transforming content creation by enabling the creation of 3D models from text and images. They have a global team distributed across North America, Asia, and Oceania and are backed by venture capital firms like Sequoia and GGV, with $52 Million in funding.
As a key member of our Data Engineering team, you will: Collaborate with Data Science, Reporting, Analytics, and other engineering teams to build data pipelines, infrastructure, and tooling to support business initiatives. Oversee the design and maintenance of data pipelines and contribute to the continual enhancement of the data engineering architecture. Collaborate with the team to meet performance, scalability, and reliability goals.
PENN Entertainment, Inc. is North America’s leading provider of integrated entertainment, sports content, and casino gaming experiences.
Design, develop, test, and maintain scalable applications using modern frameworks.
Actively participate in Agile/Scrum ceremonies, contributing to planning, estimation, and continuous improvement.
Contribute to architectural design discussions, test planning, and operational excellence initiatives.
Tealium is a trusted leader in real-time Customer Data Platforms (CDP), helping organizations unify their customer data to deliver more personalized, privacy-conscious experiences. Team Tealium has team members present in nearly 20 countries worldwide, serving customers across more than 30 countries, winning together with respect and appreciation.
Design and implement large-scale distributed data processing systems.
Build robust data pipelines and infrastructure that transform complex data into actionable insights.
Architect data lakes, warehouses, and real-time streaming platforms.
Cloudbeds is transforming hospitality with an intelligently designed platform that powers properties across 150 countries. With a completely remote team of 650+ employees across 40+ countries, they are solving challenges and redefining what's possible in hospitality.