Job Description
We are seeking a highly skilled Lead Data Engineer with strong expertise in PySpark, SQL, and Python , Azure Data Factory, Synapse, Databricks and Fabric, as well as a solid understanding of ETL and data warehousing end to end principles. The ideal candidate will have a proven track record of designing, building, and maintaining scalable data pipelines in a collaborative and fast-paced environment.
Key Responsibilities: Design and develop scalable data pipelines using PySpark to support analytics and reporting needs. Write efficient SQL and Python code to transform, cleanse, and optimize large datasets. Collaborate with machine learning engineers, product managers, and developers to understand data requirements and deliver solutions. Implement and maintain robust ETL processes to integrate structured and semi-structured data from various sources. Ensure data quality, integrity, and reliability across pipelines and systems. Participate in code reviews, troubleshooting , and performance tuning. Work independently and proactively to identify and resolve data-related issues. Contribute to Azure-based data solutions , including ADF, Synapse, ADLS, and other services. Support cloud migration initiatives and DevOps practices. Provide guidance on best practices and mentor junior team members when needed.
About Bertoni
We are a multinational team of individuals who believe that, with the right knowledge and approach, technology is the answer to the challenges businesses face today.