Design and build robust, highly scalable data pipelines and lakehouse infrastructure with PySpark, Databricks, and Airflow on AWS.
Improve the data platform development experience for Engineering, Data Science, and Product by creating intuitive abstractions, self‑service tooling, and clear documentation.
Own and maintain core data pipelines and models that power internal dashboards, ML models, and customer-facing products.
Parafin aims to grow small businesses by providing them with the financial tools they need through the platforms they already sell on. They are a Series C company backed by prominent venture capitalists, with a tight-knit team of innovators from companies like Stripe, Square, and Coinbase.
Design, build, and maintain scalable ETL pipelines for large-scale data processing.
Implement data transformations and workflows using PySpark at an intermediate to advanced level.
Optimize pipelines for performance, scalability, and cost efficiency across environments.
Truelogic is a leading provider of nearshore staff augmentation services headquartered in New York. Their team of 600+ highly skilled tech professionals, based in Latin America, drives digital disruption by partnering with U.S. companies on their most impactful projects.
Define and plan the long-term strategy for the Data Platform.
Design and develop scalable distributed systems for data management.
Improve and add features to the ETL framework while maintaining SLAs.
Jobgether is a platform that connects job seekers with companies using an AI-powered matching process. It's a platform that ensures applications are reviewed quickly, objectively, and fairly against the role's core requirements.
Design and develop scalable, maintainable, and reusable software components with a strong emphasis on performance and reliability.
Collaborate with product managers to translate requirements into well-architected solutions, owning features from design through delivery
Build intuitive and extensible user experiences using modern UI frameworks, ensuring flexibility for customer-specific needs.
ServiceNow is a global market leader that brings innovative AI-enhanced technology to over 8,100 customers, including 85% of the Fortune 500. Our intelligent cloud-based platform seamlessly connects people, systems, and processes to empower organizations to find smarter, faster, and better ways to work.
Design, build, and optimize data pipelines and workflows.
Drive scalable data solutions to support business decisions.
Contribute to architectural decisions and provide technical leadership.
Jobgether is a platform that uses AI to match candidates with jobs. They focus on ensuring fair and objective reviews of applications by using AI to identify top-fitting candidates for hiring companies.
Architect, build, and operate data infrastructure that powers Tebra’s intelligent features.
Translate business requirements into software solutions that accelerate our ability to deploy AI.
Monitor data pipelines, detect anomalies, and implement automated recovery systems.
Tebra unites Kareo and PatientPop, providing a digital backbone for practice well-being, supporting both products with a shared vision for modernized care. Over 100,000 providers trust Tebra to elevate patient experience and grow their practice, building the future of well-being with compassion and humanity.
Design and implement robust, production-grade pipelines using Python, Spark SQL, and Airflow.
Lead efforts to canonicalize raw healthcare data into internal models.
Onboard new customers by integrating their raw data into internal pipelines and canonical models.
Machinify is a healthcare intelligence company delivering value, transparency, and efficiency to health plan clients. They serve over 85 health plans, including many of the top 20, representing more than 270 million lives, with an AI-powered platform and expertise.
Play a key role in designing, developing, and delivering modern data solutions that drive business insight and innovation.
Implement scalable, high-performing cloud architectures that support analytics, AI, and operational excellence.
Be responsible for technical delivery, authoring solution documentation, and ensuring data pipelines and models meet enterprise standards for performance, reliability, and cost efficiency.
3Cloud is a company where people aren’t afraid to experiment or fail. They hire people who care about the collective growth and success of the company, challenging each other to live by 3Cloud’s core values, and resulting in amazing experiences and solutions for clients and each other.
Drive end-to-end delivery of core data engineering initiatives.
Lead and mentor the Data Engineering team.
Own data ingestion and processing for live and historical datasets.
BHFT is a proprietary algorithmic trading firm managing the full trading lifecycle — from software development to designing and deploying trading strategies. We are a 230-person company with a strong technology focus, where 70% of the team are engineers and technical specialists.
Play a Sr.tech lead & architect role to build world-class data solutions and applications that power crucial business decisions throughout the organization.
Enable a world-class engineering practice, drive the approach with which we use data, develop backend systems and data models to serve the needs of insights and play an active role in building Atlassian's data-driven culture.
Maintain a high bar for operational data quality and proactively address performance, scale, complexity and security considerations.
At Atlassian, they're motivated by a common goal: to unleash the potential of every team. Their software products help teams all over the planet and their solutions are designed for all types of work. They ensure that their products and culture continue to incorporate everyone's perspectives and experience, and never discriminate based on race, religion, national origin, gender identity or expression, sexual orientation, age, or marital, veteran, or disability status.
Design, build, and operate ETL pipelines at scale.
Design data structure for data products.
Develop and operate API/tools related to data products and machine learning products.
Mercari is a company that provides a marketplace platform. They value teamwork and provide career growth opportunities as the company continues to expand.
Design, develop, and maintain a core Python ETL framework.
Develop and optimize an automated refresh pipeline orchestrated through AWS Batch, Lambda, Step Functions, and EventBridge.
Build Python integrations with external systems that are robust, testable, and reusable.
BlastPoint is a B2B data analytics startup that helps companies engage with customers more effectively by discovering insights in their data. Founded in 2016 by Carnegie Mellon Alumni, they are a tight-knit, forward-thinking team that serves diverse industries including energy, finance, retail, and transportation.
Design and implement robust data infrastructure in AWS, using Spark with Scala
Evolve our core data pipelines to efficiently scale for our massive growth
Store data in optimal engines and formats, matching your designs to our performance needs and cost factors
tvScientific is the first CTV advertising platform purpose-built for performance marketers. They leverage data and cutting-edge science to automate and optimize TV advertising to drive business outcomes. tvScientific is built by industry leaders with history in programmatic advertising, digital media, and ad verification.
Lead support of client’s Azure Data platform and Power BI Environment, including response to any escalations while helping to analyze and resolve incidents for customers environment.
Consult, develop, and advise on solutions in Microsoft Azure with tools such as Synapse, Data Factory, Databricks, Azure ML, Data Lake, Data Warehouse, and Power BI.
Consistently learn, apply, and refine skills around data engineering and data analytics.
3Cloud hires people who aren’t afraid to experiment or fail and who are willing to give direct and candid feedback. They hire people who challenge and hold each other accountable for living 3Cloud’s core values because they know that it will result in amazing experiences and solutions for clients.
Design, build, and operate scalable data pipelines using batch and real-time processing technologies.
Build data infrastructure that ingests real-time events and stores them efficiently across databases.
Establish and enforce data contracts with backend engineering teams by implementing schema management.
Fetch provides a platform where millions of people earn rewards for buying brands they love. They have received investments from SoftBank, Univision, and Hamilton Lane and partnerships ranging from challenger brands to Fortune 500 companies. Fetch fosters a people-first culture rooted in trust, accountability, and innovation.
Develop and optimize real-time data pipelines from existing data sources to support marketing science initiatives.
Structure and organize large-scale datasets to ensure high performance, scalability, and reliability.
Build and maintain production-grade data tables powering 20–50 marketing signals across ~20 million advertisers.
ItD is a consulting and software development company blending diversity, innovation, and integrity with real business results and rejecting strong hierarchy. They are a woman- and minority-led firm employing a global community with excellent benefits such as medical, dental, vision, life insurance, paid holidays, 401K, networking & career learning and development programs.
Architect and implement Databricks Lakehouse solutions for large-scale data platforms.
Design and optimize batch & streaming data pipelines using Apache Spark (PySpark/SQL).
Implement Delta Lake best practices (ACID, schema enforcement, time travel, performance tuning).
They are looking for a Databricks Architect to design and lead modern Lakehouse data platforms using Databricks. The role focuses on building scalable, high-performance data pipelines and enabling analytics and AI use cases on cloud-native data platforms.
Design, build, and maintain pipelines that power all data use cases.
Develop intuitive, performant, and scalable data models that support product features.
Pay down technical debt, improve automation, and follow best practices in data modeling.
Patreon is a media and community platform where over 300,000 creators give their biggest fans access to exclusive work and experiences. They are leaders in the space, with over $10 billion generated by creators since Patreon's inception, with a team passionate about their mission.
Build a scalable, reliable, operable and performant big data workflow platform.
Drive the usage of Freight's data model across the organization with multiple product teams.
Drive efficiency and reliability improvements through design and automation.
Uber Freight is an enterprise technology company powering intelligent logistics with end-to-end logistics applications, managed services, and an expansive carrier network. Today, the company manages nearly $20B of freight, has one of the largest networks of carriers and is backed by best-in-class investors.
Lead the end-to-end data architecture, designing and implementing data pipelines, warehouses, and lakes that handle petabyte-scale datasets.
Collaborate with product teams to enable data-driven decision-making across the organization.
Establish best practices for data quality, governance, and security while mentoring senior engineers and conducting technical reviews.
Cority is a global enterprise EHS software provider creating industry-leading technology. They have been around for over 35 years and are known for strong employee culture and client satisfaction.