Design and build robust, highly scalable data pipelines and lakehouse infrastructure with PySpark, Databricks, and Airflow on AWS.
Improve the data platform development experience for Engineering, Data Science, and Product by creating intuitive abstractions, self‑service tooling, and clear documentation.
Own and maintain core data pipelines and models that power internal dashboards, ML models, and customer-facing products.
Parafin aims to grow small businesses by providing them with the financial tools they need through the platforms they already sell on. They are a Series C company backed by prominent venture capitalists, with a tight-knit team of innovators from companies like Stripe, Square, and Coinbase.
Architect our AWS-based data warehouse and ingestion pipelines.
Transform high-volume simulation outputs into clean, trusted datasets.
Establish schema standards and data contracts with engineering.
Onebrief provides collaboration and AI-powered workflow software designed for military staffs, making them faster, smarter, and more efficient. The company, founded in 2019, values ownership and excellence, with a team spanning veterans and technologists; it has raised $320m+ from investors and is valued at $2.15B.
Collaborate with data scientists, analysts, and other stakeholders to understand data requirements and design data models and schemas that facilitate data analysis and reporting
Design, develop, and maintain scalable and efficient data pipelines and ETL processes to ingest, process, and transform large volumes of data from various sources into usable formats
Build and optimize data storage and processing systems, including data warehouses, data lakes, and big data platforms, using AWS services such as Amazon Redshift, AWS Glue, AWS EMR, AWS S3, and AWS Lambda, to enable efficient data retrieval and analysis
ATPCO is the world's primary source for air fare content, holding over 200 million fares across 160 countries. Every day, the travel industry relies on ATPCO's technology and data solutions to help millions of travelers reach their destinations efficiently. At ATPCO, they believe in flexibility, trust, and a culture where your wellbeing comes first.
Architect and sustain self-healing pipelines using Astronomer/Airflow to ensure 24/7 data availability.
Design and optimize event-driven API ingestion frameworks leveraging AWS Lambda and DLT (Data Load Tool).
Manage high-performance modeling within AWS Redshift, utilizing DBT to transform raw transactional data into high-fidelity business intelligence.
Odisea helps close the opportunity gap between Colombia and the United States by redefining nearshoring. They are building a passionate team of professionals committed to this purpose.
Identify structural weaknesses and eliminate operational fragility.
Define clear ingestion, validation, and testing standards across the platform.
Drive ambiguous initiatives from concept to production-ready outcomes.
Life360's mission is to keep people close to the ones they love. By continuing to innovate and deliver for our customers, they have become a household name and the must-have mobile-based membership for families. Life360 has more than 500 (and growing!) remote-first employees.
Own and evolve our data infrastructure, including pipelines into our data warehouse
Manage and improve cloud infrastructure and DevOps workflows
Ensure platform reliability so product and design teams aren’t pulled into backend or operational firefighting
Meridio is a remote-first company on a mission to make health benefits for small businesses simple, affordable, and accessible. As they scale smart, they’re focused on building systems that reduce complexity instead of adding it.
Design Scalable Data Architecture: Build modern, cloud-native data platforms (AWS, Snowflake, Databricks) supporting batch and streaming use cases.
Develop Efficient Data Pipelines & Models: Automate ETL/ELT workflows, optimise data models, and enable self-serve analytics and AI.
End-to-End Data Ownership: Manage ingestion, storage, processing, and delivery of structured and unstructured data.
Trustonic provides smartphone locking technology, enabling global access to devices and digital finance. They partner with mobile carriers, retailers, and financiers across 30+ countries, powering device financing solutions. They celebrate diversity and aim to do the right thing for each other, the community, and the planet.
Implement robust data infrastructure in AWS, using Spark with Scala
Evolve core data pipelines to efficiently scale for our massive growth
Store data in optimal engines and formats
tvScientific is a CTV advertising platform purpose-built for performance marketers. They leverage data and science to automate and optimize TV advertising to drive business outcomes. TvScientific is built by industry leaders with a history in programmatic advertising and has a CTV performance platform.
Manage and support a team of 6 Data Engineers, helping them focus on impactful technical initiatives as the platform scales.
Drive execution excellence, ensuring the team delivers with high velocity, quality, and reliability.
Foster a healthy and sustainable team environment by helping the team manage workload and focus on meaningful engineering work.
Bluefish believes that AI represents the next major chapter of the internet and that consumers will increasingly use AI to consume information and media online. Bluefish is building the platform that helps brands engage consumers on this new AI channel, with powerful enterprise tools to manage AI brand safety and engage consumers with thoughtful and personalized AI marketing experiences.
Create and maintain optimal data pipeline architecture
Extend our machine learning platform by designing tools that interface with cloud services
Build the infrastructure required for optimal extraction, transformation, and loading of data
NinjaHoldings aims to revolutionize how Americans interact with financial services. They have a lean and innovative team that empowers people overlooked by traditional financial institutions through digital banking and lending products.
Design, build, and optimize data pipelines to support AI and ML projects.
Integrate data from various sources to provide a unified data view for AI applications.
Implement processes to ensure data quality, consistency, and accuracy across systems.
The Tyndale Company is a leading national supplier of arc-rated flame-resistant clothing (FRC) to the energy sector. They are a family-owned business, 9x Top Workplace winner in PA and 5x winner in TX, providing a retail-style apparel experience.
Lead teams of software engineers to create market-leading solutions.
Recruit, develop, and retain high-performing engineering talent.
Collaborate with senior leaders to shape the vision and strategy.
Jobgether is a company that uses AI to help with hiring and matching candidates with jobs. They use an AI-powered matching process to ensure applications are reviewed quickly, objectively, and fairly against the role's core requirements.
Design, develop, and maintain scalable ETL/ELT pipelines for data ingestion.
Implement data quality checks, monitoring, and validation processes.
Automate manual processes into centralized and scalable solutions.
Informa TechTarget accelerates growth from R&D to ROI, informing and connecting technology buyers and sellers. They are a vibrant community of over 2000 colleagues worldwide and traded on Nasdaq as part of Informa PLC.
Build and maintain Azure Data Factory pipelines for data ingestion.
Write Python code in Databricks for data cleaning and transformation.
Monitor daily jobs and troubleshoot pipeline failures to ensure reliability.
Jobgether is a platform that helps candidates find relevant jobs through AI-powered matching. The company ensures applications are reviewed quickly, objectively, and fairly against the role's core requirements.
Design, build, and optimize scalable data pipelines using Databricks, Apache Spark, Delta Lake, and Unity Catalog.
Develop ingestion frameworks for structured and semi‑structured data from multiple enterprise sources.
Implement data governance, data quality, and security controls across the data lifecycle.
Bridgenext is a digital consulting services leader that helps clients innovate with intention and realize their digital aspirations by creating digital products, experiences, and solutions around what real people need. Our global consulting and delivery teams facilitate highly strategic digital initiatives through digital product engineering, automation, data engineering, and infrastructure modernization services.
Experience with the integration of data from multiple data sources.
Experience with various database technologies such as SQLServer, Redshift, Postgres, and RDS.
Experience designing, building, and maintaining data pipelines.
Bluelight Consulting is a leading software consultancy dedicated to designing and developing innovative technology that enhances users' lives. With a presence across the United States and Central/South America, Bluelight is in an exciting phase of expansion, continually seeking exceptional talent to join its dynamic and diverse community.
Gathering requirements for, and implementing streaming and batch pipelines
Developing and maintaining of batch data pipelines using AWS Glue, Lambda, and Python
KOHO's mission is to make financial services better for every Canadian by offering transparent financial products designed to help users spend smart, save more, and build wealth. They are a performance organization that values autonomy, high trust, and work-life integration.
Design and implement scalable, high-throughput data ingestion systems.
Build and evolve a centralized data lake using Apache Iceberg.
Provide technical leadership through mentorship, code reviews, and design discussions.
Coupa provides a total spend management platform for businesses, which uses community-generated AI to multiply margins. They have a collaborative culture driven by transparency, openness, and a shared commitment to excellence, and are expanding their impact across the globe.
Design, build, and maintain secure, compliant ML infrastructure and automation adapted for high-sensitivity environments.
Develop and productionize machine learning and data pipelines serving real-time models that fight fraudulent traffic, spam, and bots.
Extract valuable signals from massive datasets, using your expertise to turn raw data into actionable insights.
Yelp is driven by their values, they’re a cooperative team that values individual authenticity and encourages creative solutions to problems. They are all about helping their users, growing as engineers, and having fun in a collaborative environment and are an equal opportunity employer.
Architect, build, and operate data infrastructure that powers Tebra’s intelligent features.
Translate business requirements into software solutions that accelerate our ability to deploy AI.
Monitor data pipelines, detect anomalies, and implement automated recovery systems.
Tebra unites Kareo and PatientPop, providing a digital backbone for practice well-being, supporting both products with a shared vision for modernized care. Over 100,000 providers trust Tebra to elevate patient experience and grow their practice, building the future of well-being with compassion and humanity.