Job Description
Mactores is seeking a Spark Engineer to architect, design, and build scalable data pipelines and distributed applications using Apache Spark (Spark SQL, DataFrames, RDDs). You will develop and manage ETL/ELT pipelines to process structured and unstructured data at scale, write high-performance code in Scala or PySpark for distributed data processing workloads, and optimize Spark jobs by tuning shuffle, caching, partitioning, memory, executor cores, and cluster resource allocation. The ideal candidate will monitor and troubleshoot Spark job failures, debug production issues using logs, metrics, and execution plans, and implement best practices around data quality, data reliability, security, and observability. You will support cluster provisioning, configuration, and workload optimization on platforms like Kubernetes, YARN, or EMR/Databricks.
About Mactores
Mactores is a trusted leader among businesses in providing modern data platform solutions, enabling businesses to accelerate their value through automation.