Collaborate with data scientists, analysts, and other stakeholders to understand data requirements and design data models and schemas that facilitate data analysis and reporting
Design, develop, and maintain scalable and efficient data pipelines and ETL processes to ingest, process, and transform large volumes of data from various sources into usable formats
Build and optimize data storage and processing systems, including data warehouses, data lakes, and big data platforms, using AWS services such as Amazon Redshift, AWS Glue, AWS EMR, AWS S3, and AWS Lambda, to enable efficient data retrieval and analysis
ATPCO is the world's primary source for air fare content, holding over 200 million fares across 160 countries. Every day, the travel industry relies on ATPCO's technology and data solutions to help millions of travelers reach their destinations efficiently. At ATPCO, they believe in flexibility, trust, and a culture where your wellbeing comes first.
Design, build, and maintain AWS-based data pipelines that process vehicle telemetry and validation data
Develop Python services and workflows supporting safety and performance metric computation
Interpret truck-generated data and translate it into measurable system signals
Torc is developing Level 4 autonomous semi-truck software to transform how freight moves across the world. As a part of the Daimler family with over a decade of experience, they are focused on developing software for automated trucks. Torc's culture is collaborative, energetic, and team-focused.
Design and develop high‑performance data converters for multi‑sensor autonomous‑driving data.
Design, build, and optimize large‑scale ingestion and transformation pipelines capable of processing petabyte‑scale autonomous‑driving sensor data.
Implement automated data validation, quality checks, and lineage tracking to ensure reliability of production datasets.
Torc has been a leader in autonomous driving since 2007 and is now part of the Daimler family. They are focused solely on developing software for automated trucks to transform how the world moves freight and have a collaborative, energetic, and team-focused culture.
Build pipelines to load data from various systems into Dataiku via S3 or Snowflake.
Increase the robustness of existing production pipelines, identify bottlenecks, and set up a robust monitoring, testing processes, and documentation templates.
Build custom applications and integrations to automate manual tasks related to customer operations to help Product Operations / Support / SRE in their day-to-day activities
Dataiku is the Platform for AI Success, the enterprise orchestration layer for building, deploying, and governing AI. The world’s leading companies rely on Dataiku to operationalize AI and run it as a true business performance engine delivering measurable value.
Own the delivery of scalable internal data solutions.
Translate business needs into clear technical designs and working systems.
Build and improve data pipelines, integrations, and automation.
Transparent Hiring is recruiting for a fast-growing reinsurance company operating across Germany and the United States. The environment is collaborative and driven by a strong “build and ship” mindset.
Lead the development of robust data pipelines and optimize data architecture.
Translate complex requirements into scalable data solutions.
JBS is an equal opportunity employer that values its employees. They are committed to hiring individuals authorized for employment in the United States on a W2 basis.
Gathering requirements for, and implementing streaming and batch pipelines
Developing and maintaining of batch data pipelines using AWS Glue, Lambda, and Python
KOHO's mission is to make financial services better for every Canadian by offering transparent financial products designed to help users spend smart, save more, and build wealth. They are a performance organization that values autonomy, high trust, and work-life integration.
Architect and sustain self-healing pipelines using Astronomer/Airflow to ensure 24/7 data availability.
Design and optimize event-driven API ingestion frameworks leveraging AWS Lambda and DLT (Data Load Tool).
Manage high-performance modeling within AWS Redshift, utilizing DBT to transform raw transactional data into high-fidelity business intelligence.
Odisea helps close the opportunity gap between Colombia and the United States by redefining nearshoring. They are building a passionate team of professionals committed to this purpose.
Build tools that interact with external health data networks.
Collect information about our patients and load it into Zus data stores at high volume.
Work on RESTful services that are used both internally and externally.
Zus is a shared health data platform designed to accelerate healthcare data interoperability. Founded in 2021 by Jonathan Bush, the company partners with HIEs and other data networks to aggregate patient clinical history and then translates that data into user-friendly information.
Develop and maintain backend services of our EV platform, ensuring seamless component integration.
Integrate with third-party APIs to implement external services, enhancing OEM partners and other external solutions.
Develop services to interact with connected vehicles at scale and maintain front end web apps for UIs.
EnergyHub empowers utilities and their customers to create a clean, distributed energy future. Their platform helps consumers turn smart thermostats, EVs, and other products into virtual power plants, partnering with leading IoT companies to provide energy and cost savings.
Design, build, and maintain scalable batch and real-time data pipelines that power analytics, experimentation, and machine learning
Partner cross-functionally with analytics, product, engineering and operations to deliver high-quality data solutions that drive measurable business impact
Champion data quality, reliability, and observability by implementing best practices in testing, monitoring, lineage, and incident response
Gopuff is reimagining how people purchase everyday essentials, from snacks to household goods to alcohol, all delivered in minutes. They are assembling a team of thinkers, dreamers and risk-takers who know the value of peace of mind in an unpredictable world.
Design, implement, and maintain scalable, high-performance data architectures connecting relational and non-relational systems.
Manage end-to-end data pipelines, ensuring seamless ingestion from scrapers to AI/ML workflows.
Audit and optimize existing workflows for efficiency, accuracy, and flexibility.
Jobgether is a pioneering HR Tech startup operating entirely remotely, and leading the revolution in the world of work. They are a job search engine designed exclusively for remote workers, with a team of 30 individuals located across the globe.
Create innovative solutions for handling peta-bytes of data with billions of rows & joins.
Create real time and offline features generation pipelines to managing our data infrastructure to be reliable and fast!
Develop and productionize data pipelines for our ML models in both bare-metal and the cloud environment.
Kayzen is a mobile demand-side platform (DSP) dedicated to democratizing programmatic advertising. They enable leading apps, agencies, media buyers, and brands to run programmatic customer acquisition, retargeting, and brand performance campaigns through their self-serve and managed service options.
Experience with the integration of data from multiple data sources.
Experience with various database technologies such as SQLServer, Redshift, Postgres, and RDS.
Experience designing, building, and maintaining data pipelines.
Bluelight Consulting is a leading software consultancy dedicated to designing and developing innovative technology that enhances users' lives. With a presence across the United States and Central/South America, Bluelight is in an exciting phase of expansion, continually seeking exceptional talent to join its dynamic and diverse community.
Create and maintain optimal data pipeline architecture
Extend our machine learning platform by designing tools that interface with cloud services
Build the infrastructure required for optimal extraction, transformation, and loading of data
NinjaHoldings aims to revolutionize how Americans interact with financial services. They have a lean and innovative team that empowers people overlooked by traditional financial institutions through digital banking and lending products.
Design Scalable Data Architecture: Build modern, cloud-native data platforms (AWS, Snowflake, Databricks) supporting batch and streaming use cases.
Develop Efficient Data Pipelines & Models: Automate ETL/ELT workflows, optimise data models, and enable self-serve analytics and AI.
End-to-End Data Ownership: Manage ingestion, storage, processing, and delivery of structured and unstructured data.
Trustonic makes smartphones affordable, enabling global access to devices and digital finance through secure smartphone locking technology. They partner with mobile carriers, retailers, and financiers across 30+ countries, powering device financing solutions. The company celebrates its diversity and is looking to do the right thing: for each other, the community and the planet.
Implement robust data infrastructure in AWS, using Spark with Scala
Evolve core data pipelines to efficiently scale for our massive growth
Store data in optimal engines and formats
tvScientific is a CTV advertising platform purpose-built for performance marketers. They leverage data and science to automate and optimize TV advertising to drive business outcomes. TvScientific is built by industry leaders with a history in programmatic advertising and has a CTV performance platform.
Responsible for building core infrastructure software (pipelines, APIs, data modelling) as part of our client's data platform team.
Coach & mentor other engineers to support the growth of their technical expertise.
Implementing the appropriate technologies for scaling data access patterns, batch processing, and data streaming for soft real-time consumption.
YLD is a software engineering and design consultancy that creates digital capabilities for their clients. The company has offices in London, Lisbon, and Porto and aims to attract, inspire, develop, and retain extraordinary people.
Design, develop, and maintain scalable ETL/ELT pipelines for data ingestion.
Implement data quality checks, monitoring, and validation processes.
Automate manual processes into centralized and scalable solutions.
Informa TechTarget accelerates growth from R&D to ROI, informing and connecting technology buyers and sellers. They are a vibrant community of over 2000 colleagues worldwide and traded on Nasdaq as part of Informa PLC.
Develop and maintain data models for core package application and reporting databases.
Monitor execution and performance of daily pipelines, triage and escalate any issues.
Collaborate with analytics and business teams to improve data models and data pipelines.
Bluelight Consulting designs and develops innovative software to enhance users' lives, focusing on quality and customer satisfaction. They foster a collaborative work environment where team members can grow, and are expanding across the US and Central/South America, seeking exceptional talent.