Design, build, and maintain scalable data infrastructure to support analytics and reporting across the organization.
Develop and operate ETL pipelines to ingest, transform, and deliver large-scale datasets.
Partner closely with Data Analysts and cross-functional stakeholders to provide reliable datasets and guide them in using data effectively.
Truelogic is a leading provider of nearshore staff augmentation services headquartered in New York. With over two decades of experience, they deliver top-tier technology solutions to companies of all sizes. Their team of 600+ highly skilled tech professionals, based in Latin America, drives digital disruption by partnering with U.S. companies in their projects.
Design, build, and maintain data pipelines using Snowflake, Airflow, and DBT
Lead architectural discussions around the modern data stack
Develop scalable ETL and ELT processes using Python and SQL
They are a well-funded healthcare technology company using AI and modern data infrastructure to transform how healthcare and public health decisions are made. The team is small, mission-driven, and building systems that turn raw healthcare data into actionable intelligence at scale.
Create and maintain optimal data pipeline architecture.
Assemble large, complex data sets that meet functional and non-functional business requirements.
Identify, design, and implement internal process improvements, automating manual processes, optimizing data delivery, and re-designing infrastructure for greater scalability.
Coderoad is a software development company that provides end-to-end software development services. It provides an opportunity to work on exciting, real-world projects in a supportive environment, offering staff augmentation, dedicated IT teams, and general software engineering.
You will join a team of talented engineers working closely with Data Scientists to build and scale our next-generation Ad EnGage data pipeline.
You will work with large-scale datasets (hundreds of TBs to petabyte-scale systems) using a modern data stack centered on AWS, Airflow, dbt, and Snowflake.
You’ll contribute to building reliable, high-quality data pipelines and improving the performance, scalability, and observability of our data platform.
EDO is the TV outcomes company. Their leading measurement platform connects convergent TV airings to the ad-driven consumer behaviors most predictive of future sales. They are headquartered in New York City and Los Angeles with an office space in San Francisco and recognize the benefits of hybrid working.
Design, build, and maintain scalable data pipelines for clients across industries.
Architect and optimize cloud data warehouse solutions, adapting to each client's stack.
Collaborate with analysts and data scientists to ensure data is clean, reliable, and well-modeled.
NuView Analytics helps companies accelerate the time to insights from their data through data analytics, diligence, and fractional data science. They are a growth-stage company looking to drive additional value from the data they are sitting on and value humility, intellectual rigor, and stewardship.
Organize and structure data systems at both macro and micro levels, designing and implementing data architectures that support business goalsOptimize data pipelines for performance, reliability, and scalability
Design, build, and maintain scalable ETL/ELT pipelines with Airflow to process large-scale, complex datasets
Demonstrate ability to delivery of of data products useful for machine learning and AI research and development (data models, metadata and semantics)
Owkin is an AI company on a mission to solve the complexity of biology. It is building the first Biology Super Intelligence (BASI) by combining powerful biological large language models, multimodal patient data, and agentic software.
Build pipelines to load data from various systems into Dataiku via S3 or Snowflake.
Increase the robustness of existing production pipelines, identify bottlenecks, and set up a robust monitoring, testing processes, and documentation templates.
Build custom applications and integrations to automate manual tasks related to customer operations to help Product Operations / Support / SRE in their day-to-day activities
Dataiku is the Platform for AI Success, the enterprise orchestration layer for building, deploying, and governing AI. The world’s leading companies rely on Dataiku to operationalize AI and run it as a true business performance engine delivering measurable value.
Maintain and continuously improve your technical expertise to be an Airflow expert.
Work with customers to educate and guide them regarding Airflow best practices.
Collaborate with team members to design, prototype, and implement engineering solutions.
Astronomer empowers data teams to bring software, analytics, and AI to life and is behind Astro, the unified DataOps platform powered by Apache Airflow®. They are trusted by more than 800 of the world's leading enterprises, letting businesses do more with their data.
Lead the architecture and evolution of scalable, distributed data pipelines, ensuring high availability and performance at scale
Build and maintain distributed web scraping systems using tools such as Playwright, Selenium, and BeautifulSoup
Integrate AI and LLMs into engineering workflows for code generation, automation, and optimization
MercatorAI is building scalable data infrastructure to power high-quality, data-driven decision making at scale. As an early-stage company, the team is focused on creating robust, future-ready systems that can handle complex data ingestion, transformation, and delivery across a growing national footprint.
Design, build, and maintain production data pipelines using Python, Prefect, Airflow, Jenkins or any other orchestration framework multi-phase algorithmic workflows.
Build and optimize advanced SQL transformations in Snowflake, including window functions, CTEs, stored procedures, UDFs, and semi-structured data processing.
Build and maintain dbt models for data transformation, identity resolution, and slowly changing dimension (SCD Type 2) tracking across 80+ models and multiple pipeline stages.
Kalibri helps to redefine and rebuild the hotel industry. They are looking for passionate, energetic, and hardworking people with an entrepreneurial spirit, who dream big and challenge the status quo; their team is working on cutting-edge solutions for the industry.
Assist in building and maintaining data pipelines.
Support data cleaning, validation, and quality assurance.
Contribute to data modeling and preparation for analytics and AI.
Nimble Gravity helps organizations transform and propel growth with data. They are a team of outdoor enthusiasts, adrenaline seekers, and experienced growth hackers.
Develop engineering expertise within the Dataiku Platform to help maintain and develop system integrations, platform automations, and platform configurations.
Build & maintain python & SQL data replication & data pipelines on large & often complex data sets.
Identify opportunities for improvements & optimization for greater scalability & delivery velocity
Dataiku is the Platform for AI Success, the enterprise orchestration layer for building, deploying, and governing AI. The world’s leading companies rely on Dataiku to operationalize AI and run it as a true business performance engine delivering measurable value.
Building and maintaining production-grade data pipelines in cloud data warehouses.
Designing and developing dbt models across bronze, silver, and gold layers.
Crafting easy-to-understand visualizations and dashboards in Looker or equivalent BI tools.
Plume is a trans-founded, mission-driven company with a vision to transform healthcare for every trans life by making gender-affirming hormone therapy easily accessible. They offer an affirming, trans-centered, culturally inclusive, and fun work environment filled with purpose.
Design, build, and maintain efficient data pipelines (ETL processes) to integrate data from various source systems into the data warehouse.
Develop and optimize data warehouse schemas and tables to support analytics and reporting needs.
Write and refine complex SQL queries and use scripting (e.g., Python) to transform and aggregate large datasets.
Deel is an all-in-one payroll and HR platform tailored for global teams. As one of the largest globally distributed companies, Deel's 7,000 team members span over 100 countries, fostering a dynamic culture of continuous learning and innovation.
Partner closely with business stakeholders to understand their challenges and design end-to-end architecture.
Design, develop, and own robust, efficient, and scalable data models in Snowflake and Iceberg using dbt and advanced SQL.
Build and manage reliable data pipelines and CI/CD workflows using tools like Airflow, Python, and Terraform.
Motive empowers people who run physical operations with tools to make their work safer, more productive, and more profitable. Motive serves nearly 100,000 customers and provides complete visibility and control across a wide range of industries.
Prepare and manage pre-stage files for backbook conversion activities.
Support and execute data ingestion tasks in alignment with scheduled project events, including key mock events.
Monitor and ensure data ingestion completion within defined SLA windows.
Kunai builds full-stack technology solutions for banks, credit and payment networks, infrastructure providers, and their customers. They help their clients modernize, capitalize on emerging trends, and evolve their business for the coming decades by remaining tech-agnostic and human-centered.
Lead the technical onboarding of partner institutions onto UDTS.
Design, build, and maintain scalable data pipelines and architectures.
Collaborate with team members to set engineering standards and guide data infrastructure strategy.
DataKind is a non-profit organization that uses data science and AI to address global challenges. They work with various sectors like health, humanitarian action, climate, economic opportunity, and education to create data-driven tools.
Design, build, and maintain data pipelines (ETL/ELT) in batch and streaming environments.
Develop solutions for ingesting and processing large volumes of structured, unstructured, and semi-structured data.
Create data products that respond to the analytical needs of the business.
EX Squared LATAM builds high-impact digital solutions, working with exceptional talent throughout Latin America. They foster a culture of collaboration, continuous learning, and technical excellence.
Extend, optimize, and maintain core data models for reports, machine learning, and generative AI.
Implement automation and operationalize ML models to streamline operational processes and improve efficiency.
Partner with engineering, product, and analytics teams to deliver seamless integrations and customer-facing data products.
Boulevard provides a client experience platform for appointment-based, self-care businesses, helping customers enhance client experiences. They value diversity and inclusivity, offering equal opportunities and aiming to create a supportive work environment.
Own organizational-wide data architecture, defining standards and designs.
Design and develop data pipelines, integrations, and platform features.
Partner with product managers to define new data features and capabilities.
They offer a connected equipment platform for managing mixed assets. The company values quality, continuous learning, and collaboration within a dynamic team environment.