Design, develop, and maintain data solutions, focusing on ETL processes, data quality, and data pipelines while mentoring junior developers and collaborating with stakeholders.
Project Role : Data Engineer
Project Role Description : Design, develop and maintain data solutions for data generation, collection, and processing. Create data pipelines, ensure data quality, and implement ETL (extract, transform and load) processes to migrate and deploy data across systems.
Must have skills : PySpark
Good to have skills : NA
Minimum 5 year(s) of experience is required
Educational Qualification : 15 years full time education
Summary: As a Data Engineer, you will design, develop, and maintain data solutions that facilitate data generation, collection, and processing. Your typical day will involve creating data pipelines, ensuring data quality, and implementing ETL processes to effectively migrate and deploy data across various systems, contributing to the overall efficiency and reliability of data operations. Roles and Responsibilities: Lead Development and Implementation of ETL Processes Using PYSpark: Design, develop, and optimize ETL pipelines for data ingestion, transformation, and loading into the data warehouse Design and Maintain Data Warehouse Solutions: Develop and maintain data warehouse schemas, tables, and views to support analytics and reporting needs Provide Technical Guidance and Mentorship to Junior Developers: Guide and support team members in their technical development, ensuring adherence to best practices and coding standards Collaborate with Stakeholders to Understand Data Requirements: Work closely with business users and other stakeholders to understand their needs and ensure data solutions meet their requirements Troubleshoot and Resolve Issues: Diagnose and resolve issues related to data processing, data quality, and data warehouse performance Contribute to the Development of Data Governance Policies: Help define and implement data governance policies to ensure data quality, security, and compliance Skills and Experience:Extensive Experience with PYSpark: Proficiency in Python programming, Spark architecture, Spark SQL, and Spark Data Frames Strong Knowledge of Data Warehousing Principles: Experience with data modeling, data warehousing architectures, and common data warehouse platforms (e.g., Snowflake, Redshift, BigQuery) SQL Proficiency: Strong SQL skills, including experience with relational databases and data modeling Experience with Cloud Platforms: Familiarity with cloud data warehousing services (e.g., AWS, Azure, Google Cloud) Leadership and Communication Skills: Ability to lead and motivate a team, communicate technical concepts clearly, and collaborate effectively with stakeholders Problem-Solving and Analytical Skills: Ability to diagnose and resolve complex technical issues Proven Ability to Deliver Data Solutions: Experience in designing, implementing, and maintaining data solutions Minimum Experience: At least 8 years of experience in data engineering or a related field. Professional & Technical Skills: - Must To Have Skills: Proficiency in PySpark. - Strong understanding of data pipeline architecture and design. - Experience with ETL processes and data integration techniques. - Familiarity with data quality frameworks and best practices. - Knowledge of cloud platforms and their data services. Additional Information: - The candidate should have minimum 5 years of experience in PySpark. - This position is based at our Pune office. - A 15 years full time education is required.15 years full time education
Project Role Description : Design, develop and maintain data solutions for data generation, collection, and processing. Create data pipelines, ensure data quality, and implement ETL (extract, transform and load) processes to migrate and deploy data across systems.
Must have skills : PySpark
Good to have skills : NA
Minimum 5 year(s) of experience is required
Educational Qualification : 15 years full time education
Summary: As a Data Engineer, you will design, develop, and maintain data solutions that facilitate data generation, collection, and processing. Your typical day will involve creating data pipelines, ensuring data quality, and implementing ETL processes to effectively migrate and deploy data across various systems, contributing to the overall efficiency and reliability of data operations. Roles and Responsibilities: Lead Development and Implementation of ETL Processes Using PYSpark: Design, develop, and optimize ETL pipelines for data ingestion, transformation, and loading into the data warehouse Design and Maintain Data Warehouse Solutions: Develop and maintain data warehouse schemas, tables, and views to support analytics and reporting needs Provide Technical Guidance and Mentorship to Junior Developers: Guide and support team members in their technical development, ensuring adherence to best practices and coding standards Collaborate with Stakeholders to Understand Data Requirements: Work closely with business users and other stakeholders to understand their needs and ensure data solutions meet their requirements Troubleshoot and Resolve Issues: Diagnose and resolve issues related to data processing, data quality, and data warehouse performance Contribute to the Development of Data Governance Policies: Help define and implement data governance policies to ensure data quality, security, and compliance Skills and Experience:Extensive Experience with PYSpark: Proficiency in Python programming, Spark architecture, Spark SQL, and Spark Data Frames Strong Knowledge of Data Warehousing Principles: Experience with data modeling, data warehousing architectures, and common data warehouse platforms (e.g., Snowflake, Redshift, BigQuery) SQL Proficiency: Strong SQL skills, including experience with relational databases and data modeling Experience with Cloud Platforms: Familiarity with cloud data warehousing services (e.g., AWS, Azure, Google Cloud) Leadership and Communication Skills: Ability to lead and motivate a team, communicate technical concepts clearly, and collaborate effectively with stakeholders Problem-Solving and Analytical Skills: Ability to diagnose and resolve complex technical issues Proven Ability to Deliver Data Solutions: Experience in designing, implementing, and maintaining data solutions Minimum Experience: At least 8 years of experience in data engineering or a related field. Professional & Technical Skills: - Must To Have Skills: Proficiency in PySpark. - Strong understanding of data pipeline architecture and design. - Experience with ETL processes and data integration techniques. - Familiarity with data quality frameworks and best practices. - Knowledge of cloud platforms and their data services. Additional Information: - The candidate should have minimum 5 years of experience in PySpark. - This position is based at our Pune office. - A 15 years full time education is required.15 years full time education
About Accenture
We work with one shared purpose: to deliver on the promise of technology and human ingenuity. Every day, more than 775,000 of us help our stakeholders continuously reinvent. Together, we drive positive change and deliver value to our clients, partners, shareholders, communities, and each other.We believe that delivering value requires innovation, and innovation thrives in an inclusive and diverse environment. We actively foster a workplace free from bias, where everyone feels a sense of belonging and is respected and empowered to do their best work.At Accenture, we see well-being holistically, supporting our people’s physical, mental, and financial health. We also provide opportunities to keep skills relevant through certifications, learning, and diverse work experiences. We’re proud to be consistently recognized as one of the World’s Best Workplaces™.Join Accenture to work at the heart of change.Visit us at www.accenture.com
Equal Employment Opportunity Statement
We believe that no one should be discriminated against because of their differences. All employment decisions shall be made without regard to age, race, creed, color, religion, sex, national origin, ancestry, disability status, military veteran status, sexual orientation, gender identity or expression, genetic information, marital status, citizenship status or any other basis as protected by applicable law. Our rich diversity makes us more innovative, more competitive, and more creative, which helps us better serve our clients and our communities.
Top Skills
AWS
Azure
BigQuery
GCP
Pyspark
Python
Redshift
Snowflake
Spark Architecture
Spark Sql
SQL
Accenture Pune, Mahārāshtra, IND Office
Building B-1, Magarpatta City (SEZ, Mundhwa Rd, Magarpatta, Hadapsar, Pune, Maharashtra, India, 411013
Similar Jobs
Blockchain • Fintech • Payments • Consulting • Cryptocurrency • Cybersecurity • Quantum Computing
The Sr Data Engineer will develop ML solutions using large datasets, guide teams, conduct analyses, ensure compliance, and mentor junior developers.
Top Skills:
AWSAzureGCPHadoopHivePower BIPythonPyTorchSparkSQLTableauTensorFlow
Information Technology
The Data Engineer will design and maintain data solutions, focusing on ETL processes using PySpark, ensuring data quality and optimizing data pipelines.
Top Skills:
AWSAzureBigQueryGCPPysparkRedshiftSnowflake
Information Technology
As a Data Engineer, design, develop, and maintain data solutions. Create data pipelines, ensure data quality, and implement ETL processes.
Top Skills:
Pyspark
What you need to know about the Pune Tech Scene
Once a far-out concept, AI is now a tangible force reshaping industries and economies worldwide. While its adoption will automate some roles, AI has created more jobs than it has displaced, with an expected 97 million new roles to be created in the coming years. This is especially true in cities like Pune, which is emerging as a hub for companies eager to leverage this technology to develop solutions that simplify and improve lives in sectors such as education, healthcare, finance, e-commerce and more.