Want to professionalize your AI skills, pivot to an AI role and increase your salary?
Master AI Engineering with the most practical and comprehensive LLM Development certifications at Towards AI Academy.

IBM

Senior Data Engineer - Data Platform (AWS)

IBM

Published 26 Mar 2026
Italy& Other locations
Full Time

Share this job

Role Highlights

Languages used

Python
SQL

Key skills

Data Engineer
Integrations
Critical Thinking
Cloud Security
Batch
Streaming
SAAS
ETL
ELT
IAM
Agile
Logging

Tools, Libraries and Frameworks

IOS
Red Hat
AWS S3
IBM
Kinesis
SAP
Docker
ECS
EKS
Kubernetes
PySpark

Description

\\\\Introduction\\\\ A career in IBM Consulting is rooted by long-term relationships and close collaboration with clients across the globe. You'll work with visionaries across multiple industries to modernize their data and cloud landscapes, accelerating adoption of hybrid cloud and AI-ready platforms. Your ability to drive meaningful change for clients is supported by our ecosystem of strategic partners and our technology platforms across the IBM portfolio; including Software and Red Hat. Curiosity and a constant quest for knowledge serve as the foundation to success in IBM Consulting. In your role, you'll be encouraged to challenge the norm, investigate ideas outside of your role, and come up with creative solutions resulting in measurable impact for a wide network of clients. Our culture of evolution and empathy centers on long-term career growth and development opportunities in an environment that embraces your unique skills and experience. \\\\Your role and responsibilities\\\\ We are looking for motivated and tech-passionate individuals to: Lead the development and maintenance of data ingestion pipelines and data platform components on AWS Implement data lake architectures leveraging Amazon S3, AWS Glue, AWS Lake Formation and modern table formats (e.g., Iceberg) Operate batch, streaming, and change data capture (CDC) ingestion patterns using services such as Amazon DMS, Amazon Kinesis Data Streams and Kinesis Data Firehose Integrate data from SaaS platforms and enterprise systems (e.g., SAP Datasphere) using Amazon AppFlow Develop and optimize ETL/ELT transformations in Python, PySpark and SQL across Bronze/Silver/Gold data layers Manage data cataloging, schema evolution and permission models through AWS Glue Data Catalog and AWS Lake Formation Collaborate with architects and platform leads to troubleshoot complex issues, optimize performance and resource usage, and ensure secure data operations Document technical implementations, operational procedures and best practices to support delivery teams and stakeholders \\\\Required technical and professional expertise\\\\ Minimum 3-5 years of experience in data engineering and/or cloud data workloads Strong hands-on experience in AWS analytics services such as AWS Glue, AWS Lake Formation, and Amazon S3 Experience building ingestion pipelines using Amazon DMS (full load + CDC) Familiarity with streaming ingestion using Amazon Kinesis Data Streams and Kinesis Data Firehose Proficiency in ETL/ELT development using Python, PySpark and SQL Knowledge of modern data lake and lakehouse patterns including Iceberg, partitioning strategies, and data lifecycle management Experience implementing multi-layer data models (Bronze/Silver/Gold) Experience managing data cataloging and permission models via AWS Glue Data Catalog and Lake Formation Exposure to cloud security, IAM and cost-awareness in data workloads Interest in pursuing AWS certifications aligned to data engineering or analytics \\\\Preferred technical and professional experience\\\\ Agile mindset - willingness to learn, adapt to changing priorities, take initiative, and apply critical thinking One of the following: Experience integrating SaaS platforms or SAP systems via Amazon AppFlow or similar tools Familiarity with hybrid enterprise integration scenarios (e.g., SAP Datasphere) Experience with observability tooling for data platforms (logging, metrics, tracing) Exposure to containerization/orchestration (Docker, ECS, EKS, Kubernetes) IBM is committed to creating a diverse environment and is proud to be an equal-opportunity employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, gender, gender identity or expression, sexual orientation, national origin, caste, genetics, pregnancy, disability, neurodivergence, age, veteran status, or other characteristics. IBM is also committed to compliance with all fair employment practices regarding citizenship and immigration status.

Required Qualifications and Skills

The role requires a minimum of 3-5 years of experience in data engineering and/or cloud data workloads. Strong hands-on experience with AWS analytics services such as AWS Glue, AWS Lake Formation, and Amazon S3 is necessary. Experience building ingestion pipelines using Amazon DMS, including full load and CDC, is also required. Proficiency in ETL/ELT development using Python, PySpark, and SQL is essential, along with knowledge of modern data lake and lakehouse patterns. Experience implementing multi-layer data models and managing data cataloging and permission models are key qualifications. An interest in pursuing AWS certifications aligned to data engineering or analytics is also mentioned.

Disclaimer

Disclaimer: Job and company description information and some of the data fields may have been generated via GPT-4 summarisation and could contain inaccuracies. The full external job listing link should always be relied on for authoritative information.

About the company

IBM

Size

305978

Website

ibm.com

HQ

Armonk, New York, US

Public/Private

Public Company

Description

IBM infuses core business operations with intelligence, from machine learning to generative AI, to make organizations more responsive, productive, and resilient. It helps clients put AI into action now, creating real value with trust, speed, and confidence across various areas like digital labor, IT automation, and security. The ability to utilize all data is critical, as AI's effectiveness is dependent on the quality of data fueling it, with IBM's AI, and data platform aiming to scale and accelerate AI's impact with trusted data. IBM's hybrid cloud platform offers a comprehensive approach to development, security, and operations across hybrid environments, laying a flexible foundation for leveraging data wherever it resides.

Share

Share this job

Related jobs

AI
Deep Learning
NLP
Machine Learning
Cambridge, MA, USA
Full Time
Data Engineer
Integrations
Open Source
Data Processing
Bangalore, India
Full Time
Tech Lead
API
CICD
Product Development
Pune, India
Full Time