Lead architecture and implementation of enterprise ETL/ELT pipelines, data warehouses, and performance-optimized Spark/Delta Lake workflows on Azure. Collaborate with stakeholders to deliver analytics-ready datasets, enforce data governance and compliance, monitor pipeline health, and mentor junior engineers.
Requisition Number: 2338789
Optum is a global organization that delivers care, aided by technology to help millions of people live healthier lives. The work you do with our team will directly improve health outcomes by connecting people with the care, pharmacy benefits, data and resources they need to feel their best. Here, you will find a culture guided by inclusion, talented peers, comprehensive benefits and career development opportunities. Come make an impact on the communities we serve as you help us advance health optimization on a global scale. Join us to start Caring. Connecting. Growing together.
Primary Responsibilities:
Required Qualifications:
At UnitedHealth Group, our mission is to help people live healthier lives and make the health system work better for everyone. We believe everyone - of every race, gender, sexuality, age, location and income - deserves the opportunity to live their healthiest life. Today, however, there are still far too many barriers to good health which are disproportionately experienced by people of color, historically marginalized groups and those with lower incomes. We are committed to mitigating our impact on the environment and enabling and delivering equitable care that addresses health disparities and improves health outcomes - an enterprise priority reflected in our mission.
Optum is a global organization that delivers care, aided by technology to help millions of people live healthier lives. The work you do with our team will directly improve health outcomes by connecting people with the care, pharmacy benefits, data and resources they need to feel their best. Here, you will find a culture guided by inclusion, talented peers, comprehensive benefits and career development opportunities. Come make an impact on the communities we serve as you help us advance health optimization on a global scale. Join us to start Caring. Connecting. Growing together.
Primary Responsibilities:
- Architect, design, and implement complex ETL/ELT pipelines using Azure Data Factory, Databricks Azure Function App(Python), and Mongo DB to support large-scale analytics initiatives
- Lead the development and maintenance of enterprise-grade data models and data warehouses (including star and snowflake schemas) for business intelligence and reporting
- Optimize performance of distributed data transformations leveraging Spark (Databricks), Delta Lake, partitioning strategies, caching mechanisms, broadcast joins, and resource management
- Collaborate with business stakeholders - including data scientists and analysts - to deliver robust datasets that enable machine learning modeling & advanced analytics
- Establish standards for data governance; ensure compliance with organizational security policies as well as regulatory requirements such as GDPR or HIPAA
- Monitor end-to-end workflow health using Azure Monitor/Log Analytics; proactively troubleshoot issues to maximize system uptime & reliability
- Mentor junior team members through code reviews, technical leadership sessions, documentation best practices & cross-team knowledge sharing
- Functions may include database architecture, engineering, design, optimization, security, and administration; as well as data modeling, big data development, Extract, Transform, and Load (ETL) development, storage engineering, data warehousing, data provisioning and other similar roles
- Responsibilities may include Platform-as-a-Service and Cloud solution with a focus on data stores and associated eco systems
- Duties may include management of design services, providing sizing and configuration assistance, ensuring strict data quality, and performing needs assessments
- Analyzes current business practices, processes and procedures as well as identifying future business opportunities for leveraging data storage and retrieval system capabilities
- Manage relationships with software and hardware vendors to understand the potential architectural impact of different vendor strategies and data acquisition
- May design schemas, write SQL or other data markup scripting and help to support development of Analytics and Applications that build on top of data
- Selects, develops and evaluates personnel to ensure the efficient operation of the function
- Generally, work is self-directed and not prescribed
- Works with less structured, more complex issues
- Serves as a resource for others
- Comply with the terms and conditions of the employment contract, company policies and procedures, and any and all directives (such as, but not limited to, transfer and/or re-assignment to different work locations, change in teams and/or work shifts, policies in regard to flexibility of work benefits and/or work environment, alternative work arrangements, and other decisions that may arise due to the changing business environment). The Company may adopt, vary or rescind these policies and directives in its absolute discretion and without any limitation (implied or otherwise) on its ability to do so
Required Qualifications:
- Undergraduate degree or equivalent experience
- 10+ years professional experience in enterprise-scale data engineering roles focused on cloud platforms (Azure)
- Hands-on experience with Mongo DB: schema design/development; query optimization; integration with other cloud platforms
- Experience implementing robust version control systems (Git) integrated with Databricks/Snowflake workflows
- Expertise with Azure Data Factory, Databricks notebooks (Python/PySpark/Scala), Synapse Analytics & Data Lake Storage Gen2
- Advanced proficiency in SQL; solid Python programming skills; experience optimizing Spark workloads at scale
- Proven ability to architect star/snowflake schema models for large analytical datasets
- Deep understanding of performance tuning - partitioning strategies, caching layers & broadcast joins - in distributed computing environments
At UnitedHealth Group, our mission is to help people live healthier lives and make the health system work better for everyone. We believe everyone - of every race, gender, sexuality, age, location and income - deserves the opportunity to live their healthiest life. Today, however, there are still far too many barriers to good health which are disproportionately experienced by people of color, historically marginalized groups and those with lower incomes. We are committed to mitigating our impact on the environment and enabling and delivering equitable care that addresses health disparities and improves health outcomes - an enterprise priority reflected in our mission.
Top Skills
Azure Data Factory,Databricks,Databricks Notebooks,Azure Function App (Python),Mongodb,Spark,Delta Lake,Azure Monitor,Log Analytics,Git,Snowflake,Synapse Analytics,Data Lake Storage Gen2,Python,Pyspark,Scala,Sql
Similar Jobs at Optum
Artificial Intelligence • Big Data • Healthtech • Information Technology • Machine Learning • Software • Analytics
Lead design, build, and operation of reliable batch and streaming data pipelines and AI training/evaluation datasets. Implement data quality, observability, governance for PII/PHI, and performance/cost optimizations. Partner with ML engineers on embedding/vector pipelines and RAG/LLM systems, define data contracts, create reusable frameworks, measure impact, and mentor engineers while contributing to platform architecture decisions.
Top Skills:
Sql,Python,Scala,Java,Snowflake,Databricks,Delta,Unity,Apache Airflow,Azure Data Factory,Data Lake Flow,Azure,Terraform,Bicep,Great Expectations,Dbt,Embedding Pipelines,Vector Stores,Llm,Rag
Artificial Intelligence • Big Data • Healthtech • Information Technology • Machine Learning • Software • Analytics
Maintain and secure server operating systems for a large, distributed estate. Perform patching, vulnerability and risk management, hardening and monitoring, server administration, documentation, and 24x7 on-call support while collaborating with engineering teams.
Top Skills:
Windows Server,Vmware,Esx,Microsoft Msi,Scripting,Antivirus,Aws,Azure,Gcp
Artificial Intelligence • Big Data • Healthtech • Information Technology • Machine Learning • Software • Analytics
Lead an agile release train, drive program-level cadence and ceremonies, coordinate cross-team dependencies, coach Scrum Masters and teams, maintain program metrics, remove impediments, and advance Agile/SAFe practices across the organization.
Top Skills:
Agile BoardsCa Rally (Agile Central)ConfluenceJIRASafeScrumScrum-Of-Scrums
What you need to know about the Pune Tech Scene
Once a far-out concept, AI is now a tangible force reshaping industries and economies worldwide. While its adoption will automate some roles, AI has created more jobs than it has displaced, with an expected 97 million new roles to be created in the coming years. This is especially true in cities like Pune, which is emerging as a hub for companies eager to leverage this technology to develop solutions that simplify and improve lives in sectors such as education, healthcare, finance, e-commerce and more.

