The project
About the Role
We are seeking a senior Data Architect to design and build an Enterprise Data Warehouse / Lakehouse platform for the supply chain domain from the ground up. This role requires deep expertise in Databricks Lakehouse, strong AI/ML capabilities, and a proven ability to leverage AI as an accelerator across the Software Development Life Cycle (SDLC).
Key Responsibilities
β’ Define the end-to-end vision, architecture, and roadmap for the enterprise data platform.
β’ Design and implement a greenfield EDW / Lakehouse architecture (Bronze, Silver, Gold layers).
β’ Build scalable data ingestion frameworks integrating ERP, WMS, TMS, MES, procurement, and logistics systems.
β’ Develop pipelines using PySpark, Spark SQL, Delta Live Tables (DLT), and orchestrate with Databricks Workflows.
β’ Implement data governance, lineage, and security using Unity Catalog.
β’ Embed AI/ML and GenAI capabilities such as LLM-driven schema mapping and natural-language BI solutions.
β’ Drive AI-accelerated SDLC delivery using copilots, agents, and automation tools.
β’ Establish standards for data quality, monitoring, and reliability.
Requirements
β’ 12+ years in Data Engineering / Architecture
β’ Hands-on experience with Databricks Lakehouse Platform
β’ Expertise in PySpark, Spark SQL, Delta Lake & Delta Live Tables, Databricks Workflows
β’ Experience with Unity Catalog (governance, lineage, access control)
β’ Strong knowledge of Dimensional Modeling & Data Vault, Medallion Architecture, Data Product Design
β’ Experience with CI/CD & IaC (Terraform / Databricks Asset Bundles)
β’ Solid exposure to AI/ML technologies: GenAI, LLMs, RAG, Vector Search, MLOps & Feature Stores
β’ Proven experience using AI tools (Copilots/Agents) to accelerate SDLC