Data Engineer (5+ Years Experience) – Heavy Data Analytics Project
Tech Stack: Spark, PySpark, Scala, Python, SQL, Databricks, Data Lake, Data Warehouse, Snowflake, Azure (ADF/Synapse/ADLS)
About the Role
We are hiring a Data Engineer with strong hands-on experience in building high‑performance data pipelines for a heavy data analytics project. The candidate must be excellent at writing complex aggregations, understanding business processes and analytical requirements, and designing scalable data lake and data warehouse solutions. Experience across multiple data platforms (Databricks, Snowflake, Azure Data Factory, Synapse, etc.) is a strong advantage.
Key Responsibilities
1. Data Pipeline & ETL/ELT Development
Develop, optimize, and productionize Spark (PySpark/Scala) pipelines.
Ingest, transform, cleanse, and aggregate large datasets from varied sources.
Implement scalable ETL/ELT logic for batch and near-real-time pipelines.
Apply best practices in partitioning, caching, Delta Lake optimization, and performance tuning.