The Data Engineer will build and maintain ETL/ELT pipelines, implement automated data quality checks, optimize data storage, and create interactive visualizations. They will manage CI/CD pipelines and deliver production-ready data solutions.
- Years of experience: 8 to 10 years
Focus: Reliability, Automation, and Data Quality. The Engineer turns the architect's blueprint into a working, automated reality.
- Key Responsibilities:
- Build and maintain ETL/ELT pipelines to move data from source to storage.
- Implement automated data quality testing and observability alerts.
- Optimize data storage formats (e.g., Parquet, Delta Lake) for high-speed querying.
- Manage CI/CD pipelines for data code deployment.
- Create interactive visualizations with actionable insights by persona
- Deliverables: Production-ready optimized data pipelines, Cleaned datasets, interactive & intelligent dashboards.
- Tech Stack - AWS
- Data Ingestion & Processing : AWS API Gateway, AWS Lambda, AWS Glue ETL, AWS Glue Crawler
- Data Storage & Analytics : Amazon S3, Amazon Redshift (Data Warehouse), Amazon Athena
- Governance & Security : AWS Lake Formation, AWS IAM, AWS CloudTrail, Amazon CloudWatch
- AI & Analytics : AWS Bedrock
- Visualization : Amazon QuickSight
- Visa Requirement: Preferred USA B1 Business Visa with a minimum validity of 2 years
Top Skills
Amazon Athena
Amazon Cloudwatch
Amazon Quicksight
Amazon Redshift
Amazon S3
Aws Api Gateway
Aws Bedrock
Aws Cloudtrail
Aws Glue Crawler
Aws Glue Etl
Aws Iam
Aws Lake Formation
Aws Lambda
Similar Jobs
Fintech • Professional Services • Consulting • Energy • Financial Services • Cybersecurity • Generative AI
The Data Engineer will work on GCP services and handle big data through ETL processes, employing strong problem-solving skills and analytical abilities.
Top Skills:
BigtableCloud FunctionsCloud RunDatastoreGCPJavaPysparkPythonSpannerSQL
Cloud • Fintech • Information Technology • Machine Learning • Software • App development • Generative AI
Lead architecture, design, development, and delivery of enterprise data platform components. Build and optimize ETL/ELT pipelines, data models, and warehouses (Snowflake). Mentor engineers, drive data governance, ensure performance, security, and scalability, and collaborate cross-functionally to adopt best practices and new tools.
Top Skills:
Cloud-Native (Google Cloud Preferred)CtesData WarehouseElt/Etl ToolsJavaNoSQLPythonRdbmsSnowflakeSQLStored ProceduresUdfs
Fintech • Professional Services • Consulting • Energy • Financial Services • Cybersecurity • Generative AI
Design, build, and operate scalable ETL/ELT pipelines using PySpark and AWS data services. Orchestrate workflows with Apache Airflow, implement AWS Glue jobs and Data Catalog, manage Lake Formation permissions, publish datasets for BI, and deliver QuickSight visualizations while ensuring data quality and performance.
Top Skills:
Pyspark,Apache Airflow,Aws Glue,Aws Lake Formation,Aws Glue Data Catalog,Amazon Quicksight
What you need to know about the Pune Tech Scene
Once a far-out concept, AI is now a tangible force reshaping industries and economies worldwide. While its adoption will automate some roles, AI has created more jobs than it has displaced, with an expected 97 million new roles to be created in the coming years. This is especially true in cities like Pune, which is emerging as a hub for companies eager to leverage this technology to develop solutions that simplify and improve lives in sectors such as education, healthcare, finance, e-commerce and more.


