Mastercard
Lead Data Engineer - (Datawarehouse) - Apache Nifi, Python, PySpark, Hadoop, Cloudera platforms, and Airflow
Be an Early Applicant
Lead the development of high-quality, scalable data pipelines and manage data warehousing projects in a Big Data environment. Optimize performance and troubleshoot issues in Spark applications, while collaborating in Agile teams to enhance data workflows and standards.
Our Purpose
Mastercard powers economies and empowers people in 200+ countries and territories worldwide. Together with our customers, we're helping build a sustainable economy where everyone can prosper. We support a wide range of digital payments choices, making transactions secure, simple, smart and accessible. Our technology and innovation, partnerships and networks combine to deliver a unique set of products and services that help people, businesses and governments realize their greatest potential.
Title and Summary
Lead Data Engineer - (Datawarehouse) - Apache Nifi, Python, PySpark, Hadoop, Cloudera platforms, and Airflow
Job Description Summary
Lead Data Engineer (Datawarehouse)
Job Posting Title
Who is Mastercard?
Mastercard is a global technology company in the payments industry. Our mission is to connect and power an inclusive, digital economy that benefits everyone, everywhere by making transactions safe, simple, smart, and accessible. Using secure data and networks, partnerships and passion, our innovations and solutions help individuals, financial institutions, governments, and businesses realize their greatest potential.
Our decency quotient, or DQ, drives our culture and everything we do inside and outside of our company. With connections across more than 210 countries and territories, we are building a sustainable world that unlocks priceless possibilities for all.
Overview
Mastercard is looking for a Lead Software Engineer to join the Account Level Management(ALM) team in our Pune office that's focused on building ALM Services with Data warehouse skills. The Mastercard Account Level Management platform empowers real-time card level decisioning. As consumers progress along their life stages as card holders, with increasing disposable income and more refined preferences, ALM provides services to issuers so they can effectively offer more relevant benefits and rewards at each stage, to drive loyalty and spend.
Skill Requirements:
MUST Technical• 8-12 of overall career experience years of experience in Data Warehouse• Expertise in Python, PySpark, Hadoop, Cloudera platforms, and Airflow• Solid Experience of building complex data pipelines through PySpark with Scala/Python on Hadoop or Object storage• Experience of building Nifi pipelines (Preferred).• Proficiency in Cloudera platforms, and Airflow• Expertise in Data Engineering and implementing multiple end-to-end DW projects in Big Data environment handling petabyte scale data.• Experience of working with Databases like Oracle, Netezza and have strong SQL knowledge.• Proficient in working within an Agile/Scrum framework, including creating user stories with well-defined acceptance criteria, participating in sprint planning and reviews• Write and maintain Unix shell scripts, Oracle SQL, PL/SQL, and perform SQL tuning.• Optimize and troubleshoot Spark applications for performance, scalability, and fault tolerance• Use Git-based version control systems and CI/CD pipelines (e.g., Jenkins).• Implement and manage HIVE external tables, partitions, and various file formats.• Work across on-premises and cloud environments (AWS, Azure, Databricks).• Strong experience with Hadoop ecosystem and Cloudera Data Platform (CDP).
Optional Technical• Strong analytical skills required for debugging production issues, providing root cause and implementing mitigation plan• Strong communication skills - both verbal and written • Ability to multi-task across multiple projects, interface with external / internal resources• Proactive, detail-oriented and able to function under pressure in an independent environment along with a high degree of initiative and self-motivation to drive results • Willingness to quickly learn and implement new technologies, participate POC to explore best solution for the problem statement• Experience working diverse and geographically distributed project teams
• Education:• Bachelor's degree in information systems, Information Technology, Computer Science or Engineering or equivalent work experience.
Role
o Develop high quality, secure and scalable data pipelines using spark, Scala/Python/Java on Hadoop or object storage like MinIO.
o Leverage technologies and solutions to innovate with increasingly large data sets.
o Drive automation and efficiency in Data ingestion, data movement and data access workflows by innovation and collaboration.
o Understand, implement and enforce Software development standards and engineering principles in the Big Data space.
o Contribute ideas to help ensure that required standards and processes are in place and actively look for opportunities to enhance standards and improve process efficiency.
o Perform assigned tasks and production incident independently.
Corporate Security Responsibility
All activities involving access to Mastercard assets, information, and networks comes with an inherent risk to the organization and, therefore, it is expected that every person working for, or on behalf of, Mastercard is responsible for information security and must:
Mastercard powers economies and empowers people in 200+ countries and territories worldwide. Together with our customers, we're helping build a sustainable economy where everyone can prosper. We support a wide range of digital payments choices, making transactions secure, simple, smart and accessible. Our technology and innovation, partnerships and networks combine to deliver a unique set of products and services that help people, businesses and governments realize their greatest potential.
Title and Summary
Lead Data Engineer - (Datawarehouse) - Apache Nifi, Python, PySpark, Hadoop, Cloudera platforms, and Airflow
Job Description Summary
Lead Data Engineer (Datawarehouse)
Job Posting Title
Who is Mastercard?
Mastercard is a global technology company in the payments industry. Our mission is to connect and power an inclusive, digital economy that benefits everyone, everywhere by making transactions safe, simple, smart, and accessible. Using secure data and networks, partnerships and passion, our innovations and solutions help individuals, financial institutions, governments, and businesses realize their greatest potential.
Our decency quotient, or DQ, drives our culture and everything we do inside and outside of our company. With connections across more than 210 countries and territories, we are building a sustainable world that unlocks priceless possibilities for all.
Overview
Mastercard is looking for a Lead Software Engineer to join the Account Level Management(ALM) team in our Pune office that's focused on building ALM Services with Data warehouse skills. The Mastercard Account Level Management platform empowers real-time card level decisioning. As consumers progress along their life stages as card holders, with increasing disposable income and more refined preferences, ALM provides services to issuers so they can effectively offer more relevant benefits and rewards at each stage, to drive loyalty and spend.
Skill Requirements:
MUST Technical• 8-12 of overall career experience years of experience in Data Warehouse• Expertise in Python, PySpark, Hadoop, Cloudera platforms, and Airflow• Solid Experience of building complex data pipelines through PySpark with Scala/Python on Hadoop or Object storage• Experience of building Nifi pipelines (Preferred).• Proficiency in Cloudera platforms, and Airflow• Expertise in Data Engineering and implementing multiple end-to-end DW projects in Big Data environment handling petabyte scale data.• Experience of working with Databases like Oracle, Netezza and have strong SQL knowledge.• Proficient in working within an Agile/Scrum framework, including creating user stories with well-defined acceptance criteria, participating in sprint planning and reviews• Write and maintain Unix shell scripts, Oracle SQL, PL/SQL, and perform SQL tuning.• Optimize and troubleshoot Spark applications for performance, scalability, and fault tolerance• Use Git-based version control systems and CI/CD pipelines (e.g., Jenkins).• Implement and manage HIVE external tables, partitions, and various file formats.• Work across on-premises and cloud environments (AWS, Azure, Databricks).• Strong experience with Hadoop ecosystem and Cloudera Data Platform (CDP).
Optional Technical• Strong analytical skills required for debugging production issues, providing root cause and implementing mitigation plan• Strong communication skills - both verbal and written • Ability to multi-task across multiple projects, interface with external / internal resources• Proactive, detail-oriented and able to function under pressure in an independent environment along with a high degree of initiative and self-motivation to drive results • Willingness to quickly learn and implement new technologies, participate POC to explore best solution for the problem statement• Experience working diverse and geographically distributed project teams
• Education:• Bachelor's degree in information systems, Information Technology, Computer Science or Engineering or equivalent work experience.
Role
o Develop high quality, secure and scalable data pipelines using spark, Scala/Python/Java on Hadoop or object storage like MinIO.
o Leverage technologies and solutions to innovate with increasingly large data sets.
o Drive automation and efficiency in Data ingestion, data movement and data access workflows by innovation and collaboration.
o Understand, implement and enforce Software development standards and engineering principles in the Big Data space.
o Contribute ideas to help ensure that required standards and processes are in place and actively look for opportunities to enhance standards and improve process efficiency.
o Perform assigned tasks and production incident independently.
Corporate Security Responsibility
All activities involving access to Mastercard assets, information, and networks comes with an inherent risk to the organization and, therefore, it is expected that every person working for, or on behalf of, Mastercard is responsible for information security and must:
- Abide by Mastercard's security policies and practices;
- Ensure the confidentiality and integrity of the information being accessed;
- Report any suspected information security violation or breach, and
- Complete all periodic mandatory security trainings in accordance with Mastercard's guidelines.
Top Skills
Airflow
Apache Nifi
AWS
Azure
Cloudera
Databricks
Git
Hadoop
Netezza
Oracle
Pyspark
Python
SQL
Unix
Mastercard Pune, Mahārāshtra, IND Office



Poona Club Road, Pune, Maharashtra, India, 411001
Similar Jobs at Mastercard
Blockchain • Fintech • Payments • Consulting • Cryptocurrency • Cybersecurity • Quantum Computing
The Software Engineer II will develop data-driven solutions for sustainability, focusing on energy consumption and emission reduction data pipelines and APIs. Responsibilities include data model engineering, ensuring data integrity, and collaborating across teams to improve systems and insights.
Top Skills:
Cloud-Native TechnologiesPower AppsSharepointSplunkSQL
Blockchain • Fintech • Payments • Consulting • Cryptocurrency • Cybersecurity • Quantum Computing
The Lead Platform Engineer will support the NaaS ecosystem, ensuring system stability and automating network operations while delivering tooling improvements.
Top Skills:
AnsibleGithub ActionsGitlab CiGoJenkinsReactTerraform
Blockchain • Fintech • Payments • Consulting • Cryptocurrency • Cybersecurity • Quantum Computing
The Senior Software Engineer will develop enterprise applications using Java and front-end technologies, ensuring compliance and performance while supporting business needs and architecture design.
Top Skills:
AngularAWSC SharpChefCSSEc2EcsGradleHibernateHTMLJ2EeJavaJavaScriptJenkinsJqueryJspMavenNoSQLReactS3ServletSonarSpring BootSQLUnixXMLXslt
What you need to know about the Pune Tech Scene
Once a far-out concept, AI is now a tangible force reshaping industries and economies worldwide. While its adoption will automate some roles, AI has created more jobs than it has displaced, with an expected 97 million new roles to be created in the coming years. This is especially true in cities like Pune, which is emerging as a hub for companies eager to leverage this technology to develop solutions that simplify and improve lives in sectors such as education, healthcare, finance, e-commerce and more.




