Job Description

We are seeking a highly skilled Lead Data Engineer with strong expertise in PySpark, SQL, and Python, as well as a solid understanding of ETL and data warehousing principles. The ideal candidate will have a proven track record of designing, building, and maintaining scalable data pipelines in a collaborative and fast-paced environment. Key Responsibilities: Design and develop scalable data pipelines using PySpark to support analytics and reporting needs. Write efficient SQL and Python code to transform, cleanse, and optimize large datasets. Collaborate with machine learning engineers, product managers, and developers to understand data requirements and deliver solutions. Implement and maintain robust ETL processes to integrate structured and semi-structured data from various sources. Ensure data quality, integrity, and reliability across pipelines and systems. Participate in code reviews, troubleshooting, and performance tuning. Work independently and proactively to identify and resolve data-related issues. If applicable, contribute to Azure-based data solutions , including ADF, Synapse, ADLS, and other services. Support cloud migration initiatives and DevOps practices, if relevant to the role. Provide guidance on best practices and mentor junior team members when needed.

About Bertoni Solutions

We are a multinational team that believes technology can solve business challenges with the right knowledge. Since 2016, we've helped clients translate technology into success.

Apply for This Position