You will play a key role in building and maintaining scalable data infrastructure within a cloud-first environment. Your primary focus will be on designing and implementing data solutions using Azure Databricks, ensuring efficient processing and integration across diverse data sources.
Key Responsibilities
- Develop and manage data workflows using Azure Databricks and Azure Data Factory
- Write and optimize transformation logic in PySpark and SQL for high-performance processing
- Design and maintain Lakehouse patterns using Delta Lake for reliable data storage and access
- Build robust ETL and ELT processes to support analytics and reporting needs
- Integrate heterogeneous data sources into centralized data platforms
- Design and evolve data warehouse models to support business intelligence initiatives
- Ensure data accuracy, system scalability, and processing efficiency at scale
- Collaborate with analytics teams to enable effective reporting in Power BI
- Maintain and enhance existing SQL Server and SSIS-based data systems as needed
- Contribute to the evolution of cloud-native data architecture strategies
Technology Environment
Core tools include Azure Databricks, Spark, PySpark, Delta Lake, Azure Data Factory, SQL Server, Power BI, and SSIS. You’ll work across cloud and hybrid environments, focusing on modernization and performance.