Remote Data Jobs · PySpark

Job listings

Design, develop and optimize ETL pipelines using PySpark, Hive and Airflow for large-scale data processing. Build and maintain data lakes using Delta Lake, ensuring data reliability, quality, and integrity. Monitor, troubleshoot, and improve pipeline performance, scalability, and reliability.

We are looking for a highly skilled Data Engineer to join our team and contribute to the design, development, and optimization of data-driven financial solutions. The ideal candidate will have strong technical expertise and a collaborative mindset. You will develop and maintain robust data pipelines and scalable processing solutions, work with AWS services and collaborate with cross-functional teams.

Seeking an experienced and dynamic Azure Data Engineer to lead data engineering engagements, design scalable data solutions, and help clients unlock the full potential of their data. The candidate will act as a trusted advisor, providing technical leadership and strategic guidance while working on cutting-edge Microsoft Azure technologies.

$130,000–$140,000/yr

Looking for a Data Engineer II to help design, build, and continuously improve the GoGuardian Analytics and AI/ML ecosystem. This position sits on the Data Engineering team, a group responsible for building and maintaining the core data platform that powers analytics, product insights, and machine learning across the company. Collaborate closely with Data Science, Business Intelligence, and other teams to enable the next generation of data-driven products and AI capabilities.

Integrate with our team as a Data Engineer and contribute to building scalable and intelligent solutions for large volumes of data. If you are passionate about data, have solid experience with Databricks, Spark, Azure Data Factory (ADF), Cloudera and data modeling, enjoy solving complex problems and collaborating with multidisciplinary teams, this opportunity is for you!

Mactores is seeking a highly skilled and innovative Spark Engineer to design, develop, optimize, and operationalize high-performance data pipelines and applications using Apache Spark. This role requires hands-on expertise in distributed data processing, ETL engineering, performance tuning, and cluster management.

$77,500–$107,000/yr

This role builds data pipelines that enable analytics and modeling across strategic initiatives including plant measurement and environmental controls. Day-to-day tasks include monitoring of data processing, troubleshooting and debugging data processing tasks, and development of new data flows for existing platforms. This role supports the development of data processing pipelines for Sensei Ag.

We are looking for a talented and motivated person to act as a Data Developer, integrating with our team and contributing to the construction of scalable and intelligent solutions for large volumes of data. Responsibilities include designing, developing, and maintaining scalable and robust data pipelines, creating ingestion, transformation, and data modeling solutions using Databricks, Spark/PySpark, Cloudera, and Azure Data Factory (ADF).

Estamos em busca de uma pessoa Senior Data Engineer para atuar em um projeto internacional , na construção e evolução do ecossistema de dados, garantindo escalabilidade, qualidade e governança em todas as camadas. You will develop and maintain robust and scalable data pipelines and build and evolve data models using good engineering practices.