Architect and implement Databricks Lakehouse solutions for large-scale data platforms.
Design and optimize batch & streaming data pipelines using Apache Spark (PySpark/SQL).
Implement Delta Lake best practices (ACID, schema enforcement, time travel, performance tuning).
They are looking for a Databricks Architect to design and lead modern Lakehouse data platforms using Databricks. The role focuses on building scalable, high-performance data pipelines and enabling analytics and AI use cases on cloud-native data platforms.
Lead support of client’s Azure Data platform and Power BI Environment, including response to any escalations while helping to analyze and resolve incidents for customers environment.
Consult, develop, and advise on solutions in Microsoft Azure with tools such as Synapse, Data Factory, Databricks, Azure ML, Data Lake, Data Warehouse, and Power BI.
Consistently learn, apply, and refine skills around data engineering and data analytics.
3Cloud hires people who aren’t afraid to experiment or fail and who are willing to give direct and candid feedback. They hire people who challenge and hold each other accountable for living 3Cloud’s core values because they know that it will result in amazing experiences and solutions for clients.
Design and implement data pipelines using Databricks, PySpark, and Delta Lake.
Work closely with business stakeholders and analysts to understand KPIs.
Model and structure data using dimensional modeling techniques.
Clear Tech specializes in Data, Analytics, and Artificial Intelligence, helping companies around the world transform their data into real business value. Our team combines highly skilled talent in Latin America with global best practices across cloud technologies and delivers end-to-end projects.
Play a key role in designing, developing, and delivering modern data solutions that drive business insight and innovation.
Implement scalable, high-performing cloud architectures that support analytics, AI, and operational excellence.
Be responsible for technical delivery, authoring solution documentation, and ensuring data pipelines and models meet enterprise standards for performance, reliability, and cost efficiency.
3Cloud is a company where people aren’t afraid to experiment or fail. They hire people who care about the collective growth and success of the company, challenging each other to live by 3Cloud’s core values, and resulting in amazing experiences and solutions for clients and each other.
Design and build robust, highly scalable data pipelines and lakehouse infrastructure with PySpark, Databricks, and Airflow on AWS.
Improve the data platform development experience for Engineering, Data Science, and Product by creating intuitive abstractions, self‑service tooling, and clear documentation.
Own and maintain core data pipelines and models that power internal dashboards, ML models, and customer-facing products.
Parafin aims to grow small businesses by providing them with the financial tools they need through the platforms they already sell on. They are a Series C company backed by prominent venture capitalists, with a tight-knit team of innovators from companies like Stripe, Square, and Coinbase.
Build and scale distributed data pipelines for large-scale time series, log data, and high-volume event streams.
Design and maintain reliable, high-performance Spark and Python workflows to support model training datasets.
Analyze and resolve performance bottlenecks related to latency, memory utilization, data skew, and throughput.
ItD blends diversity, innovation, and integrity with real business results as a consulting and software development company. Their structure rejects any strong hierarchy, empowering them to deliver excellent results as a woman- and minority-led firm.
Implement robust data infrastructure in AWS, using Spark with Scala
Evolve core data pipelines to efficiently scale for our massive growth
Store data in optimal engines and formats
tvScientific is a CTV advertising platform purpose-built for performance marketers. They leverage data and science to automate and optimize TV advertising to drive business outcomes. TvScientific is built by industry leaders with a history in programmatic advertising and has a CTV performance platform.
Build and maintain Azure Data Factory pipelines for data ingestion.
Write Python code in Databricks for data cleaning and transformation.
Monitor daily jobs and troubleshoot pipeline failures to ensure reliability.
Jobgether is a platform that helps candidates find relevant jobs through AI-powered matching. The company ensures applications are reviewed quickly, objectively, and fairly against the role's core requirements.
Design, develop, and maintain a core Python ETL framework.
Develop and optimize an automated refresh pipeline orchestrated through AWS Batch, Lambda, Step Functions, and EventBridge.
Build Python integrations with external systems that are robust, testable, and reusable.
BlastPoint is a B2B data analytics startup that helps companies engage with customers more effectively by discovering insights in their data. Founded in 2016 by Carnegie Mellon Alumni, they are a tight-knit, forward-thinking team that serves diverse industries including energy, finance, retail, and transportation.
Architect and develop cloud-native data platforms, focusing on modern data warehousing, transformation, and orchestration frameworks.
Design scalable data pipelines and models, ensure data quality and observability, and contribute to backend services and infrastructure supporting data-driven features.
Collaborate across multiple teams, influence architectural decisions, mentor engineers, and implement best practices for CI/CD and pipeline delivery.
Jobgether uses an AI-powered matching process to ensure applications are reviewed quickly, objectively, and fairly against the role's core requirements. Their system identifies the top-fitting candidates, and this shortlist is then shared directly with the hiring company.
Design, build, and maintain pipelines that power all data use cases.
Develop intuitive, performant, and scalable data models that support product features.
Pay down technical debt, improve automation, and follow best practices in data modeling.
Patreon is a media and community platform where over 300,000 creators give their biggest fans access to exclusive work and experiences. They are leaders in the space, with over $10 billion generated by creators since Patreon's inception, with a team passionate about their mission.
Design and implement robust, production-grade pipelines using Python, Spark SQL, and Airflow.
Lead efforts to canonicalize raw healthcare data into internal models.
Onboard new customers by integrating their raw data into internal pipelines and canonical models.
Machinify is a healthcare intelligence company delivering value, transparency, and efficiency to health plan clients. They serve over 85 health plans, including many of the top 20, representing more than 270 million lives, with an AI-powered platform and expertise.
Architect, build, and operate data infrastructure that powers Tebra’s intelligent features.
Translate business requirements into software solutions that accelerate our ability to deploy AI.
Monitor data pipelines, detect anomalies, and implement automated recovery systems.
Tebra unites Kareo and PatientPop, providing a digital backbone for practice well-being, supporting both products with a shared vision for modernized care. Over 100,000 providers trust Tebra to elevate patient experience and grow their practice, building the future of well-being with compassion and humanity.
Design, build, and scale the lakehouse architecture that underpins analytics, machine learning, and AI.
Modernize our data ecosystem, making it discoverable, reliable, governed, and ready for self-service and intelligent automation.
Operate anywhere along the data lifecycle from ingestion and transformation to metadata, orchestration, and MLOps.
OnX is a pioneer in digital outdoor navigation with a suite of apps. With more than 400 employees, they have created regional “Basecamps” to help remote employees find connection and inspiration.
Design, build, and optimize data pipelines and workflows.
Drive scalable data solutions to support business decisions.
Contribute to architectural decisions and provide technical leadership.
Jobgether is a platform that uses AI to match candidates with jobs. They focus on ensuring fair and objective reviews of applications by using AI to identify top-fitting candidates for hiring companies.
Design, build, and maintain robust ETL/ELT pipelines to ingest large-scale datasets and high-frequency streams.
Lead the design and evolution of our enterprise data warehouse, ensuring it is scalable and performant.
Manage our data transformation layer using Dataform (preferred) or dbt to orchestrate complex, reliable workflows.
UW provides utilities all in one place, including energy, broadband, mobile, and insurance. They aim to double in size and offer savings to customers, fostering a culture that values imaginative and pragmatic problem-solvers.
Develop, implement, and sustain a comprehensive data quality framework to systematically monitor, validate, and enhance data accuracy and consistency throughout all systems.
Operationalize the enterprise data governance framework, aligning with stakeholder needs related to data quality, access controls, compliance, privacy, and security.
Partner with data engineers, architects, product teams, and analysts to define data quality requirements and ensure alignment with business objectives.
Abacus Insights aims to unlock the power of data so health plans can enable the right care at the right time—making life better for millions of people. They are backed by $100M from top VCs, and tackle big challenges in an industry that’s ready for change, mastering turning complex healthcare data into clear, actionable insights.
Design and optimize scalable data pipelines and data architecture.
Build cloud-native data solutions using Azure, Databricks (Unity Catalog & Delta Lake), and other big data technologies.
Contribute to a strong data culture through continuous learning and knowledge sharing.
Redcare Pharmacy is Europe’s No.1 e-pharmacy, powered by passionate teams and cutting-edge innovation. They strive to create a healthy collaborative work environment where every employee feels valued and inspired to contribute to their vision.
Build and maintain Azure Data Factory pipelines to ingest data from multiple sources.
Write Python code in Databricks to clean raw data and move it into the silver layer, handling deduplication, type casting, and validation.
Monitor daily jobs and troubleshoot any failures to ensure pipeline stability.
Jobgether is a platform that leverages AI to connect job seekers with employers. They focus on ensuring fair and efficient application reviews, connecting top candidates directly with hiring companies.
Design and implement scalable, high-performing data pipelines and optimize our data architecture.
Build and deploy cloud-native solutions leveraging Azure Data Services, Databricks and other Big Data technologies.
Collaborate across teams to understand and support their data needs while ensuring the data architecture supports ongoing and future initiatives.
Redcare Pharmacy is Europe’s No.1 e-pharmacy, powered by passionate teams and cutting-edge innovation. They strive to create a healthy collaborative work environment where every employee feels valued and inspired to contribute to their vision “Until every human has their health”.