Design, build, and optimize ETL and data pipelines on GCP (BigQuery, Dataflow, Cloud Storage). Develop Apache Beam Dataflow jobs, automate workflows with Airflow/Cloud Composer, integrate diverse data sources, ensure data quality, monitoring, governance, and collaborate with stakeholders and analysts to support analytics and reporting.
Title:
Senior Data Engineer / ETL Engineer
Location: Pune / Hyderabad / Mohali
Job Summary:
As a Senior Data Engineer / ETL Engineer, you will be instrumental in designing, developing, and optimizing data processing systems that support our organization's data initiatives. Your expertise in Bigdata and Google Cloud Platform (GCP) will be essential in building robust ETL pipelines that transform raw data into actionable insights. You will collaborate with cross-functional teams to ensure that data is accurate, accessible, and valuable for decision-making.
Responsibilities:
- Design and Develop ETL Processes:
- Lead the design and implementation of ETL processes using all kinds of batch/streaming tools to extract, transform, and load data from various sources into GCP.
- Collaborate with stakeholders to gather requirements and ensure that ETL solutions meet business needs.
- Data Pipeline Optimization:
- Optimize data pipelines for performance, scalability, and reliability, ensuring efficient data processing workflows.
- Monitor and troubleshoot ETL processes, proactively addressing issues and bottlenecks.
- Data Integration and Management:
- Integrate data from diverse sources, including databases, APIs, and flat files, ensuring data quality and consistency.
- Manage and maintain data storage solutions in GCP (e.g., BigQuery, Cloud Storage) to support analytics and reporting.
- GCP Dataflow Development:
- Write Apache Beam based Dataflow Job for data extraction, transformation, and analysis, ensuring optimal performance and accuracy.
- Collaborate with data analysts and data scientists to prepare data for analysis and reporting.
- Automation and Monitoring:
- Implement automation for ETL workflows using tools like Apache Airflow or Cloud Composer, enhancing efficiency and reducing manual intervention.
- Set up monitoring and alerting mechanisms to ensure the health of data pipelines and compliance with SLAs.
- Data Governance and Security:
- Apply best practices for data governance, ensuring compliance with industry regulations (e.g., GDPR, HIPAA) and internal policies.
- Collaborate with security teams to implement data protection measures and address vulnerabilities.
- Documentation and Knowledge Sharing:
- Document ETL processes, data models, and architecture to facilitate knowledge sharing and onboarding of new team members.
- Conduct training sessions and workshops to share expertise and promote best practices within the team.
Qualifications:
- Education:
- Bachelor’s degree in Computer Science, Information Systems, or a related field.
- Experience:
- Minimum of 5 years of industry experience in data engineering or ETL development, with a strong focus on Data Stage and GCP.
- Proven experience in designing and managing ETL solutions, including data modeling, data warehousing, and SQL development.
- Technical Skills:
- Strong knowledge of GCP services (e.g., BigQuery, Dataflow, Cloud Storage, Pub/Sub) and their application in data engineering.
- Experience of cloud-based solutions, especially in GCP, cloud certified candidates are preferred.
- Experience and knowledge of Big Data data processing in batch mode and streaming mode, proficient in Bigdata eco systems, e.g. Hadoop, HBase, Hive, MapReduce, Kafka, Flink, Spark, etc.
- Familiarity with Java & Python for data manipulation on Cloud/Bigdata platform.
- Analytical Skills:
- Strong problem-solving skills with a keen attention to detail.
- Ability to analyze complex data sets and derive meaningful insights.
Top Skills
Datastage,Gcp,Bigquery,Dataflow,Cloud Storage,Pub/Sub,Apache Beam,Apache Airflow,Cloud Composer,Hadoop,Hbase,Hive,Mapreduce,Kafka,Flink,Spark,Java,Python
Similar Jobs
Fintech • Professional Services • Consulting • Energy • Financial Services • Cybersecurity • Generative AI
The role involves managing finance change, regulatory reporting, project management, and software testing, requiring extensive experience in financial services and team leadership.
Top Skills:
Automation Testing ToolsExcelSQL
Blockchain • Fintech • Payments • Consulting • Cryptocurrency • Cybersecurity • Quantum Computing
Lead franchise product development by converting business requirements into engineering features, defining reusable frameworks, assessing platform architecture, and launching scalable MVPs and pilots with cross-functional teams to drive Mastercard's multi-rail strategy and commercial success.
Top Skills:
AIConfluenceFoundryMatchMcp
Blockchain • Fintech • Payments • Consulting • Cryptocurrency • Cybersecurity • Quantum Computing
Lead technical product initiatives in the payments domain: translate customer and business requirements into functional and platform solutions, write technical design documents, manage backlog and user stories, coordinate cross-functional teams, identify risks from operational performance, and communicate findings to senior leadership.
Top Skills:
Fis Connex
What you need to know about the Pune Tech Scene
Once a far-out concept, AI is now a tangible force reshaping industries and economies worldwide. While its adoption will automate some roles, AI has created more jobs than it has displaced, with an expected 97 million new roles to be created in the coming years. This is especially true in cities like Pune, which is emerging as a hub for companies eager to leverage this technology to develop solutions that simplify and improve lives in sectors such as education, healthcare, finance, e-commerce and more.

