· Lead the design and implementation of the Data Lakehouse architecture on Azure, making critical technology choices across Azure Data Lake Storage Gen2, Delta Lake, Azure Synapse, or Azure Databricks.
· Design, develop, and optimize high-throughput data processing pipelines (ETL/ELT) for a diverse range of sources (on-premise, cloud, streaming, APIs, files, etc.), ensuring resilience and fault tolerance.
· Act as a key technical liaison with AI, BI, and DevOps teams, driving the integration of data into complex analytics applications, production-grade machine learning models, and intelligent reporting systems.
· Pioneer and implement advanced optimization strategies for large-scale data storage and query systems (big data), focusing on maximizing performance, scalability, and cost-effectiveness.
· Establish and enforce best practices for data governance, including comprehensive data quality frameworks, centralized metadata management, and automated data lineage processes.
· Collaborate with infrastructure and security teams to architect and operate secure, compliant data systems, ensuring adherence to security protocols, internal regulations, and external standards.
· Mentor junior engineers, review code/design specifications, and produce high-quality technical documentation that enables knowledge transfer and efficient operations across product and analytics teams.









