Want to professionalize your AI skills, pivot to an AI role and increase your salary?
Master AI Engineering with the most practical and comprehensive LLM Development certifications at Towards AI Academy.

IBM

data architect - Application Architect-Azure Cloud Migration

IBM

Published 31 Mar 2026
Gurgaon, India
Full Time

Share this job

Role Highlights

Languages used

Python

Key skills

Data Engineer
Data Architect
Data Infrastructure
Data Lakes
Computer Science
Integrations
CICD
Data Processing
Data Warehousing
Cloud
AI
Architecture
ETL
ELT
Security
Cluster
Devops
Storage
Modelling

Tools, Libraries and Frameworks

Data Factory
Azure Databricks
IOS
Red Hat
Azure DevOps
Spark SQL
IBM
DataBricks
Kafka
RBAC
GitHub
PySpark
NoteBooks

Description

\\\\Introduction\\\\ A career in IBM Consulting is built on long-term client relationships and close collaboration worldwide. Youll work with leading companies across industries, helping them shape their hybrid cloud and AI journeys. With support from our strategic partners, robust IBM technology, and Red Hat, youll have the tools to drive meaningful change and accelerate client impact. At IBM Consulting, curiosity fuels success. Youll be encouraged to challenge the norm, explore new ideas, and create innovative solutions that deliver real results. Our culture of growth and empathy focuses on your long-term career development while valuing your unique skills and experiences. \\\\Your role and responsibilities\\\\ We are seeking a seasoned Azure Data Architect to lead our data modernization initiatives, leveraging Microsoft Fabric and the modern Azure Data Stack. The ideal candidate will design end-to-end data solutions, transforming legacy data architectures into high-performance lakehouse/mesh architectures. You will define best practices for data ingestion, processing, and governance while acting as the technical authority for our data engineering teams. Key Responsibilities 1\\. Architectural Design & Strategy (Microsoft Fabric Focus) \\\ Define and implement end-to-end data architecture strategies covering Data Lakes, Warehouses, and Lakehouses using Microsoft Fabric (OneLake, Lakehouse, DirectLake). \\\ Architect modern data integration patterns, prioritizing Fabric Dataflows Gen2 Fabric Data Pipelines, and Azure Data Factory (ADF). \\\ Design and implement Data Mesh architectures, focusing on domain-driven design and reusable data products. 2\\. Data Engineering & Pipeline Development \\\ Design high-performance, scalable ETL/ELT pipelines using PySpark SQL, and Databricks. \\\ Implement Medallion Architecture (Bronze, Silver, Gold layers) for structured data processing. \\\ Architect real-time and event-driven ingestion patterns using Fabric Event Streams KQL, and Kafka. 3\\. Governance, Security, and Optimization \\\ Establish governance frameworks for OneLake, including RBAC, data lineage, and naming conventions. \\\ Implement data cataloging, sensitivity labeling, and security using Microsoft Purview. \\\ Optimize cloud-based data infrastructure for performance, cost-effectiveness, and scalability (including cluster management in Databricks). 4\\. Leadership & Consulting \\\ Lead client workshops, requirement discovery sessions, and modern data stack roadmaps. \\\ Mentor data engineers on best practices for Spark/PySpark development and Azure DevOps CI/CD implementation. \\\ Support pre-sales activities, including effort estimation and technical proposals. \\\\Required technical and professional expertise\\\\ Required Technical Skills & Experience \\\ Azure Services: Deep expertise in Microsoft Fabric, Azure Data Factory (ADF), Azure Data Lake Storage (ADLS Gen2), Azure Databricks, and Azure Synapse Analytics. \\\ Languages: Strong proficiency in Python/PySpark and SQL. \\\ Modeling: Experience with Dimensional Modeling, Data Warehousing, and Delta Lake concepts. \\\ Data Fabric: Hands-on experience with Dataflows Gen2, OneLake, and Notebooks. \\\ DevOps: Experience in CI/CD pipelines (Azure DevOps/GitHub). \\\\Preferred technical and professional experience\\\\ \\\ Bachelors or Masters degree in Computer Science, Engineering, or a related field. \\\* Preferred Certification: Microsoft Certified: Fabric Analytics Engineer Associate (DP-700) or Azure Data Engineer Associate (DP-203) IBM is committed to creating a diverse environment and is proud to be an equal-opportunity employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, gender, gender identity or expression, sexual orientation, national origin, caste, genetics, pregnancy, disability, neurodivergence, age, veteran status, or other characteristics. IBM is also committed to compliance with all fair employment practices regarding citizenship and immigration status.

Required Qualifications and Skills

The role requires deep expertise in Azure Services including Microsoft Fabric, Azure Data Factory, Azure Data Lake Storage Gen2, Azure Databricks, and Azure Synapse Analytics. Strong proficiency in Python/PySpark and SQL is necessary, along with experience in Dimensional Modeling, Data Warehousing, and Delta Lake concepts. Hands-on experience with Data Fabric components like Dataflows Gen2, OneLake, and Notebooks is essential. Experience with CI/CD pipelines using Azure DevOps or GitHub is also required. A Bachelors or Masters degree in Computer Science, Engineering, or a related field is preferred.

Disclaimer

Disclaimer: Job and company description information and some of the data fields may have been generated via GPT-4 summarisation and could contain inaccuracies. The full external job listing link should always be relied on for authoritative information.

About the company

IBM

Size

305978

Website

ibm.com

HQ

Armonk, New York, US

Public/Private

Public Company

Description

IBM infuses core business operations with intelligence, from machine learning to generative AI, to make organizations more responsive, productive, and resilient. It helps clients put AI into action now, creating real value with trust, speed, and confidence across various areas like digital labor, IT automation, and security. The ability to utilize all data is critical, as AI's effectiveness is dependent on the quality of data fueling it, with IBM's AI, and data platform aiming to scale and accelerate AI's impact with trusted data. IBM's hybrid cloud platform offers a comprehensive approach to development, security, and operations across hybrid environments, laying a flexible foundation for leveraging data wherever it resides.

Share

Share this job

Related jobs

AI
Deep Learning
NLP
Machine Learning
Cambridge, MA, USA
Full Time
Data Engineer
Integrations
Open Source
Data Processing
Bangalore, India
Full Time
Tech Lead
API
CICD
Product Development
Pune, India
Full Time