Source Job

  • Support the development and maintenance of data pipelines using Databricks, Spark, and similar technologies.
  • Write and optimize SQL and Python scripts for data transformation, integration, and automation tasks.
  • Develop automation scripts that populate metadata and comments across Databricks tables using structured definitions.

Python SQL Cloud

20 jobs similar to Data Engineering Intern(Spring/Summer 2026)

Jobs ranked by similarity.

$69,100–$107,400/yr

  • Assist in executing data engineering projects within the Customer Intelligence portfolio to meet defined timelines and deliverables.
  • Build and maintain ETL pipelines based on user and project specifications to enable reliable data movement.
  • Develop and update technical documentation for key systems and data assets.

Stryker is one of the world’s leading medical technology companies and, together with its customers, is driven to make healthcare better.

$110,000–$140,000/yr
US

  • Design, build, and maintain scalable and reliable data pipelines.
  • Develop and maintain ETL data pipelines for large volumes of data, writing clean, maintainable, and efficient code.
  • Work closely with product managers, data scientists, and software engineers to create and prepare datasets from disparate sources.

Curinos empowers financial institutions to make better, faster and more profitable decisions through industry-leading proprietary data, technologies and insights.

$215,000–$240,000/yr
US

Own the design, build, and optimization of end-to-end data pipelines. Establish and enforce best practices in data modeling, orchestration, and system reliability. Collaborate with stakeholders to translate requirements into robust, scalable data solutions.

YipitData is the leading market research and analytics firm for the disruptive economy and most recently raised $475M from The Carlyle Group at a valuation of over $1B.

Latam

Work with data end-to-end, exploring, cleaning, and assembling large, complex datasets. Analyze raw data from multiple sources and identify trends and patterns, maintaining reliable data pipelines. Build analytics-ready outputs and models that enable self-service and trustworthy insights across the organization.

Truelogic is a leading provider of nearshore staff augmentation services headquartered in New York, delivering top-tier technology solutions for over two decades.

  • Architect and implement scalable Lakehouse solutions using Delta Tables and Delta Live Tables.
  • Design and orchestrate complex data workflows using Databricks Workflows and Jobs.
  • Develop production-grade Python and PySpark code, including custom Python libraries.

Coderio designs and delivers scalable digital solutions for global businesses with a strong technical foundation and a product mindset.

US

  • Assist in estimating effort for QA tasks under the guidance of senior team members.
  • Review requirement and mapping documents to help identify potential gaps or clarifications needed.
  • Support the creation of test scenarios and contribute to writing basic ETL test cases.

Abacus Insights is changing the way healthcare works. They unlock the power of data so health plans can enable the right care at the right time—making life better for millions of people. The company is backed by $100M from top VCs, and they value bold, curious, and collaborative employees.

Europe Unlimited PTO

Design, implement, and maintain scalable ETL/ELT pipelines using Python, SQL, and modern orchestration frameworks. Build and optimize data models and schemas for cloud warehouses and relational databases, supporting AI and analytics workflows. Lead large-scale data initiatives from planning through execution, ensuring performance, cost efficiency, and reliability.

This position is posted by Jobgether on behalf of a partner company.

US Unlimited PTO

  • Partner with clients and implementation teams to understand data distribution requirements.
  • Design and develop data pipelines integrating with Databricks and Snowflake, ensuring accuracy and integrity.
  • Lead architecture and implementation of solutions for health plan clients, optimizing cloud-based technologies.

Abacus Insights is changing the way healthcare works by unlocking the power of data to enable the right care at the right time. Backed by $100M from top VCs, they're tackling big challenges in an industry that’s ready for change with a bold, curious, and collaborative team.

$155,000–$185,000/yr
US Unlimited PTO

As a Senior Data Engineer, shape a scalable data platform that drives business insights. Design and maintain robust data pipelines and collaborate with cross-functional teams. Tackle complex data challenges, implement best practices, and mentor junior engineers.

Jobgether is a Talent Matching Platform that partners with companies worldwide to efficiently connect top talent with the right opportunities through AI-driven job matching.

Europe

  • Support production systems and help triage issues during live sporting events
  • Architect low-latency, real-time analytics systems including raw data collection, feature development and endpoint production
  • Integrate large and complex real-time datasets into new consumer and enterprise products

Swish Analytics is a sports analytics, betting, and fantasy startup building the next generation of predictive sports analytics data products. They're looking for team-oriented individuals with an authentic passion for accurate and predictive real-time data who can execute in a fast-paced, creative, and continually-evolving environment without sacrificing technical excellence.

US

  • Design and engineer robust data pipelines using technologies like Databricks, Azure Data Factory, Apache Spark, and Delta Lake.
  • Craft healthcare data solutions - processing massive healthcare datasets, optimizing performance, and ensuring data is accurate and secure.
  • Communicate technical concepts to non-technical stakeholders, manage multiple priorities, and meet deadlines.

Gentiva offers compassionate care in the comfort of patients' homes as a national leader in hospice, palliative, home health care, and advanced illness management. They have nearly 600 locations and thousands of clinicians across 38 states, offering rewarding careers in a collaborative environment.

$115,000–$160,000/yr
US

As a key member of our Data Engineering team, you will: Collaborate with Data Science, Reporting, Analytics, and other engineering teams to build data pipelines, infrastructure, and tooling to support business initiatives. Oversee the design and maintenance of data pipelines and contribute to the continual enhancement of the data engineering architecture. Collaborate with the team to meet performance, scalability, and reliability goals.

PENN Entertainment, Inc. is North America’s leading provider of integrated entertainment, sports content, and casino gaming experiences.

$78,900–$147,500/yr
US

  • Design, develop, and maintain robust data processes and solutions.
  • Develop and maintain data models, databases, and data warehouses.
  • Collaborate with stakeholders to gather requirements and provide data solutions.

Highmark Health is a national, blended health organization that includes one of America’s largest Blue Cross Blue Shield insurers.

  • Design and implement data ingestion and transformation pipelines using Databricks, PySpark, and distributed processing.
  • Implement Delta Lake principles, focusing on CDC and schema evolution, integrating data quality frameworks within CI/CD pipelines for data integrity.
  • Develop and optimize complex SQL and Python scripts, handle both structured and unstructured data, and improve inconsistent legacy datasets.

Mobile Wave Solutions is a professional services company specializing in software development as a service, with a team of over 120 engineers.

$155,000–$180,000/yr
US

  • Design, build, and maintain robust and scalable data pipelines from diverse sources.
  • Leverage expert-level experience with dbt and Snowflake to structure, transform, and organize data.
  • Collaborate with engineering, product, and analytics teams to deliver data solutions that drive business value.

Topstep is an engaging working environment which ranges from fully remote to hybrid and they foster a culture of collaboration.

  • Design, build, and maintain the pipelines that power all data use cases.
  • Develop intuitive, performant, and scalable data models that support product features, internal analytics, experimentation, and machine learning workloads.
  • Define and enforce standards for accuracy, completeness, lineage, and dependency management.

Patreon is a media and community platform where over 300,000 creators give their biggest fans access to exclusive work and experiences.

$150,000–$185,000/yr

  • Design, build, and oversee the deployment of technology for managing structured and unstructured data.
  • Develop tools leveraging AI, ML, and big-data to cleanse, organize, and transform data.
  • Design and maintain CI/CD pipelines using GitHub Actions to automate deployment, testing, and monitoring.

NBCUniversal is one of the world's leading media and entertainment companies creating world-class content across film, television, streaming, theme parks, and more.

India

  • Build and operate data pipelines from D365, Power Platform, and other sources into the enterprise data platform.
  • Design and implement star schemas, data lake house structures, and semantic models for Power BI.
  • Optimize performance and cost management for reporting in Azure.

Jobgether uses an AI-powered matching process to ensure applications are reviewed quickly and fairly. They identify top-fitting candidates and share the shortlist with the hiring company.

$130,000–$176,000/yr
US Unlimited PTO

  • Design, develop, and implement end-to-end data pipelines to support data collection and transformation.
  • Lead the architecture and development of scalable and maintainable data solutions.
  • Collaborate with data scientists and analysts to provide clean and accessible data.

DexCare optimizes time in healthcare, streamlining patient access, reducing waits, and enhancing overall experiences.

  • Design, develop, and maintain scalable data pipelines using Snowflake and dbt.
  • Write and optimize advanced SQL queries for performance and reliability.
  • Implement ETL/ELT processes to ingest and transform data from multiple sources.

Nagarro is a digital product engineering company that is scaling in a big way and builds products, services, and experiences that inspire, excite, and delight.