Collaborate with data scientists, analysts, and other stakeholders to understand data requirements and design data models and schemas that facilitate data analysis and reporting
Design, develop, and maintain scalable and efficient data pipelines and ETL processes to ingest, process, and transform large volumes of data from various sources into usable formats
Build and optimize data storage and processing systems, including data warehouses, data lakes, and big data platforms, using AWS services such as Amazon Redshift, AWS Glue, AWS EMR, AWS S3, and AWS Lambda, to enable efficient data retrieval and analysis
ATPCO is the world's primary source for air fare content, holding over 200 million fares across 160 countries. Every day, the travel industry relies on ATPCO's technology and data solutions to help millions of travelers reach their destinations efficiently. At ATPCO, they believe in flexibility, trust, and a culture where your wellbeing comes first.
Develop and maintain data models for core package application and reporting databases.
Monitor execution and performance of daily pipelines, triage and escalate any issues.
Collaborate with analytics and business teams to improve data models and data pipelines.
Bluelight Consulting designs and develops innovative software to enhance users' lives, focusing on quality and customer satisfaction. They foster a collaborative work environment where team members can grow, and are expanding across the US and Central/South America, seeking exceptional talent.
Create and maintain optimal data pipeline architecture
Extend our machine learning platform by designing tools that interface with cloud services
Build the infrastructure required for optimal extraction, transformation, and loading of data
NinjaHoldings aims to revolutionize how Americans interact with financial services. They have a lean and innovative team that empowers people overlooked by traditional financial institutions through digital banking and lending products.
Design, build, maintain, and operate scalable streaming and batch data pipelines.
Work with AWS services, including Redshift, EMR, and ECS, to support data processing and analytics workloads.
Develop and maintain data workflows using Python and SQL.
Southworks helps companies with software development and digital transformation. They focus on solving complex problems and delivering innovative solutions.
Assemble and manage large, complex datasets that meet both non-functional and functional business requirements.
Identify, design, and implement internal process improvements to enhance scalability, optimize data delivery, and automate manual processes.
Build and maintain optimal data pipeline architecture for efficient extraction, transformation, and loading of data from various sources.
Sanford Health is one of the largest and fastest-growing not-for-profit health systems in the United States, dedicated to the work of health and healing. The organization has 53,000 employees and serves over 2 million patients across the upper Midwest.
Design, build, and optimize scalable data architectures that power marketing analytics and survey measurement initiatives.
Deliver automated, high-impact data solutions and insights that enhance decision-making across teams.
Build robust pipelines, dashboards, and analytical frameworks in fast-paced environments.
ItD is a consulting and software development company that blends diversity, innovation, and integrity with real business results. They reject any strong hierarchy, empowering teams to deliver excellent results in a woman- and minority-led firm.
Gathering requirements for, and implementing streaming and batch pipelines
Developing and maintaining of batch data pipelines using AWS Glue, Lambda, and Python
KOHO's mission is to make financial services better for every Canadian by offering transparent financial products designed to help users spend smart, save more, and build wealth. They are a performance organization that values autonomy, high trust, and work-life integration.
Design, build, and optimize data pipelines to support AI and ML projects.
Integrate data from various sources to provide a unified data view for AI applications.
Implement processes to ensure data quality, consistency, and accuracy across systems.
The Tyndale Company is a leading national supplier of arc-rated flame-resistant clothing (FRC) to the energy sector. They are a family-owned business, 9x Top Workplace winner in PA and 5x winner in TX, providing a retail-style apparel experience.
Partner with BI Analysts, Operations, Product, and Engineering teams to define and assess data requirements.
Design, implement, and maintain ETL/ELT pipelines and data integrations.
Build and manage data architecture, including relational and dimensional databases or cloud data warehouses.
Jobgether is an AI-powered matching service that helps candidates get reviewed quickly, objectively, and fairly. They identify top-fitting candidates and share this shortlist directly with the hiring company.
Architect our AWS-based data warehouse and ingestion pipelines.
Transform high-volume simulation outputs into clean, trusted datasets.
Establish schema standards and data contracts with engineering.
Onebrief provides collaboration and AI-powered workflow software designed for military staffs, making them faster, smarter, and more efficient. The company, founded in 2019, values ownership and excellence, with a team spanning veterans and technologists; it has raised $320m+ from investors and is valued at $2.15B.
Lead discovery conversations to understand client goals.
Design and deliver technical roadmaps for data platform adoption.
Build modern, reliable data pipelines and ETL/ELT frameworks.
InterWorks is a tech consultancy that empowers clients with customized, collaborative solutions. They value unique contributions, and their people are the glue that holds their business together as they pursue innovation alongside people who inspire them.
Lead the development of robust data pipelines and optimize data architecture.
Translate complex requirements into scalable data solutions.
JBS is an equal opportunity employer that values its employees. They are committed to hiring individuals authorized for employment in the United States on a W2 basis.
Own the delivery of scalable internal data solutions.
Translate business needs into clear technical designs and working systems.
Build and improve data pipelines, integrations, and automation.
Transparent Hiring is recruiting for a fast-growing reinsurance company operating across Germany and the United States. The environment is collaborative and driven by a strong “build and ship” mindset.
Lead the design and implementation of scalable ETL pipelines and data lakes in AWS
Develop and optimise data architectures for terabyte-scale relational and distributed data systems
Collaborate with Data Scientists, Software Engineers, and Architects to integrate data solutions into analytics platforms and applications
Smart Working connects skilled professionals with outstanding global teams for full-time, long-term roles. They help discover meaningful work with teams that invest in your success, empowering you to grow personally and professionally in a remote-first world.
Design, build, and maintain scalable data pipelines using Microsoft Fabric and Apache Airflow
Ingest, transform, and integrate data from a variety of sources, including relational systems, APIs, and MongoDB
Design and maintain analytical data models, including fact and dimension tables, to support reporting and analytics
Theoria Medical is a comprehensive medical group and technology company dedicated to serving patients across the care continuum with an emphasis on post-acute care and primary care. Theoria serves facilities across the United States with a multitude of services to improve the quality of care delivered, refine facility processes, and enhance critical relationships.
Design, develop, and maintain ETL/ELT pipelines on cloud-based data platforms.
Build data ingestion, transformation, and orchestration workflows using tools such as Azure Data Factory, Airflow, Fivetran, or similar.
Develop transformations and data processing logic using platforms such as Databricks, Snowflake, or equivalent.
Ankura Consulting Group, LLC is an independent global expert services and advisory firm. They deliver services and end-to-end solutions to help clients at critical inflection points related to conflict, crisis, performance, risk, strategy, and transformation, and consists of more than 2000 professionals.
Design and build robust, highly scalable data pipelines and lakehouse infrastructure with PySpark, Databricks, and Airflow on AWS.
Improve the data platform development experience for Engineering, Data Science, and Product by creating intuitive abstractions, self‑service tooling, and clear documentation.
Own and maintain core data pipelines and models that power internal dashboards, ML models, and customer-facing products.
Parafin aims to grow small businesses by providing them with the financial tools they need through the platforms they already sell on. They are a Series C company backed by prominent venture capitalists, with a tight-knit team of innovators from companies like Stripe, Square, and Coinbase.
Build and maintain scalable data pipelines from ingestion through transformation and delivery.
Design, build, and maintain our data warehouse and data marts.
Partner with stakeholders to translate business needs into clean data models.
Gurobi Optimization focuses on mathematical optimization. They empower customers to expand their use of mathematical optimization technology in order to make smarter decisions and solve some of the world's toughest and most impactful business problems.
Design, develop, and maintain a core Python ETL framework.
Develop and optimize an automated refresh pipeline orchestrated through AWS Batch, Lambda, Step Functions, and EventBridge.
Build Python integrations with external systems that are robust, testable, and reusable.
BlastPoint is a B2B data analytics startup that helps companies engage with customers more effectively by discovering insights in their data. Founded in 2016 by Carnegie Mellon Alumni, they are a tight-knit, forward-thinking team that serves diverse industries including energy, finance, retail, and transportation.