The Data Architect will design scalable data architectures on Databricks, lead data migration initiatives, and define data governance frameworks.
Overview
We are seeking an experienced Data Architect with extensive expertise in designing and implementing modern data architectures. This role requires strong software engineering principles, hands-on coding abilities, and experience building data engineering frameworks. The ideal candidate will have a proven track record of implementing Databricks-based solutions in the healthcare industry, with expertise in data catalog implementation and governance frameworks.
About the Role
As a Data Architect, you will be responsible for designing and implementing scalable, secure, and efficient data architectures on the Databricks platform. You will lead the technical design of data migration initiatives from legacy systems to modern Lakehouse architecture, ensuring alignment with business requirements, industry best practices, and regulatory compliance.
Key Responsibilities
- Design and implement modern data architectures using Databricks Lakehouse platform
- Lead the technical design of Data Warehouse/Data Lake migration initiatives from legacy systems
- Develop data engineering frameworks and reusable components to accelerate delivery
- Establish CI/CD pipelines and infrastructure-as-code practices for data solutions
- Implement data catalog solutions and governance frameworks
- Create technical specifications and architecture documentation
- Provide technical leadership to data engineering teams
- Collaborate with cross-functional teams to ensure alignment of data solutions
- Evaluate and recommend technologies, tools, and approaches for data initiatives
- Ensure data architectures meet security, compliance, and performance requirements
- Mentor junior team members on data architecture best practices
- Stay current with emerging technologies and industry trends
Qualifications
- Extensive experience in data architecture design and implementation
- Strong software engineering background with expertise in Python or Scala
- Proven experience building data engineering frameworks and reusable components
- Experience implementing CI/CD pipelines for data solutions
- Expertise in infrastructure-as-code and automation
- Experience implementing data catalog solutions and governance frameworks
- Deep understanding of Databricks platform and Lakehouse architecture
- Experience migrating workloads from legacy systems to modern data platforms
- Strong knowledge of healthcare data requirements and regulations
- Experience with cloud platforms (AWS, Azure, GCP) and their data services
- Bachelor's degree in computer science, Information Systems, or related field; advanced degree preferred
Technical Skills
- Programming languages: Python and/or Scala (required)
- Data processing frameworks: Apache Spark, Delta Lake
- CI/CD tools: Jenkins, GitHub Actions, Azure DevOps
- Infrastructure-as-code (optional): Terraform, CloudFormation, Pulumi
- Data catalog tools: Databricks Unity Catalog, Collibra, Alation
- Data governance frameworks and methodologies
- Data modeling and design patterns
- API design and development
- Cloud platforms: AWS, Azure, GCP
- Container technologies: Docker, Kubernetes
- Version control systems: Git
- SQL and NoSQL databases
- Data quality and testing frameworks
Optional - Healthcare Industry Knowledge
- Healthcare data standards (HL7, FHIR, etc.)
- Clinical and operational data models
- Healthcare interoperability requirements
- Healthcare analytics use cases
Top Skills
Alation
Spark
AWS
Azure
Azure Devops
CloudFormation
Collibra
Databricks Unity Catalog
Delta Lake
Docker
GCP
Github Actions
Jenkins
Kubernetes
NoSQL
Pulumi
Python
Scala
SQL
Terraform
Similar Jobs
Artificial Intelligence • Fintech • Information Technology • Logistics • Payments • Business Intelligence • Generative AI
Lead cross-functional initiatives in program execution, manage risks, oversee project phases, and implement scalable frameworks within a SaaS environment.
Top Skills:
GitJenkinsJIRA
Cloud • Fintech • Information Technology • Machine Learning • Software • App development • Generative AI
The Senior Software Engineer will design, develop, and maintain cloud-based applications, leading innovation and mentoring other engineers in an agile environment.
Top Skills:
.Net Core.Net FrameworkAWSAzureC#GCPKafkaRabbitMQSQL
Cloud • Information Technology • Security • Software
As a Software Engineer in the Data Engineering team, you'll design and maintain data pipelines, optimize MongoDB clusters, and collaborate with teams to support data capabilities of JumpCloud's products.
Top Skills:
AWSAzureCloudFormationElkGCPGoJavaKafkaLinuxMongoDBNode.jsOsxPythonTerraform
What you need to know about the Pune Tech Scene
Once a far-out concept, AI is now a tangible force reshaping industries and economies worldwide. While its adoption will automate some roles, AI has created more jobs than it has displaced, with an expected 97 million new roles to be created in the coming years. This is especially true in cities like Pune, which is emerging as a hub for companies eager to leverage this technology to develop solutions that simplify and improve lives in sectors such as education, healthcare, finance, e-commerce and more.