Job Description
Integrate with our team as a Data Engineer and contribute to building scalable and intelligent solutions for large volumes of data. If you are passionate about data, have solid experience with Databricks, Spark, Azure Data Factory (ADF), Cloudera and data modeling, enjoy solving complex problems and collaborating with multidisciplinary teams, this opportunity is for you!
Responsibilities:
- Design, develop and maintain scalable and robust data pipelines.
- Create solutions for data ingestion, transformation and modeling using - Databricks, Spark/PySpark, Cloudera and Azure Data Factory (ADF).
- Ensure the quality, integrity and usability of data throughout the pipeline.
- Design and maintain Data Lake and Data Warehouse solutions, applying good data governance practices.
- Work with concepts and structures such as Bronze, Silver and Gold layers, Star Schema, Delta Tables, Delta Sharing and analytical tables.
- Collaborate with technical and business teams to orchestrate solutions aligned with the company's objectives.
Technical Requirements:
- Solid experience as a Data Engineer or similar role. Advanced knowledge in: - SQL - Python / PySpark - Azure Databricks (workflows, jobs, delta tables, queries, etc) - Azure Data Factory (ADF) - Cloudera (Hive, Impala, HDFS, etc) - ETL/ELT and data pipelines - Dimensional data modeling - Data Lake / Data Warehouse - Data governance and quality - Experience with CI/CD practices applied to the data environment. - Ability to communicate and collaborate with multidisciplinary teams. - Ability to work in dynamic environments and deal with a poorly defined technical scope. - Knowledge in code optimization in cloud environments.
Differentiator: - Experience with Teradata - Experience with SAS.
About CI&T
CI&T specializes in technological transformation, combining human expertise with AI to create scalable tech solutions.