Quantiphi Analytics Solution Pvt Ltd

Senior Data Architect

Quantiphi Analytics Solution Pvt Ltd
Mumbai/Bombay
Not disclosed
Work from OfficeWork from Office
Full TimeFull Time
Min. 8 yearsMin. 8 years

Job Description

Architect - Data

While technology is the heart of our business, a global and diverse culture is the heart of our success. We love our people and we take pride in catering them to a culture built on transparency, diversity, integrity, learning and growth.


If working in an environment that encourages you to innovate and excel, not just in professional but personal life, interests you- you would enjoy your career with Quantiphi!

Data Architect

Exp Range : 8 - 13 Years

Job location : Mumbai , Bangalore, Trivandrum

Role Overview

The Data Architect is the senior technical owner of the platform's design. You will define and evolve the architectural blueprint — the canonical data model, the ingestion framework, the transformation patterns, the FHIR serialization layer, the bidirectional flow with FHIR-Repository, and the governance and observability frameworks that hold them together. You will set the standards every other role implements against.

This role works in close partnership with the customer's existing Health Data Engine technical owners, who carry deep operational knowledge of healthcare data at scale. Architectural decisions are made in dialogue with them, anchored on real volume requirements and known operational pain points. You will be expected to defend choices with technical depth and adapt them when warranted.

Key Responsibilities

  • Own the end-to-end architecture across ingestion (Flink + PySpark), storage (Iceberg on GCS with BigLake Metastore), transformation (dbt over Starburst), FHIR serialization (flat FHIR Iceberg → bundles → FHIR-Repository), and consumption (Starburst, FHIR API, data products). Maintain the architecture specification document and its companion design docs.

  • Lead the design and evolution of the Common Data Model (CDM): the dimensional, fact, bridge, and reference table classes; SCD2 semantics; hash key conventions; the write-authority matrix that governs cross-source field precedence. CDM is the project's central design artifact and the architect owns it.

  • Define the bidirectional FHIR flow with origin-tag-based loop prevention. Specify the FHIR repository egress interceptor contract, the loopback Flink consumer pipeline, and the boundary semantics between FHIR-repo-owned and CDM-owned fields.

  • Define the patient identity resolution architecture using Informatica MDM, including the synchronous-call pattern at ingestion, the asynchronous ECI change event pipeline, and the DLQ taxonomy for MDM failures.

  • Set the standards every other role implements against — naming conventions, hash algorithms (SHA-256 BINARY 32, 0x1F separator, NULL handling), Iceberg table properties (CoW vs MoR, partitioning, compaction cadence), DLQ taxonomy, observability (metric naming, structured logging, correlation ID propagation), and security (PHI handling, encryption, audit logging).

  • Own the spec-driven development framework's program-level and component-level specs. Approve significant changes to platform-wide rules. Architect specs are the constitution every implementing engineer references.

  • Drive the high-volume capacity design — sustained 50K msg/sec ingest target with 150K peak, billions of rows in CDM facts, thousand-concurrent Starburst workloads, thousands per second FHIR API. Lead capacity planning, validation, and the parallel-run cutover from the existing Health Data Engine.

  • Evaluate and recommend technology choices that are still open — Confluent Cloud tier, Starburst Galaxy vs Enterprise, FHIR-Repository sizing, multi-region failover topology. Build option analyses; defend recommendations with concrete trade-off matrices.

  • Provide architectural review for engineering work. Review specs at the component and unit level when they touch architectural concerns. Mentor data engineers and data modelers on design principles.

  • Lead architectural review meetings with customer technical stakeholders — including the Health Data Engine retirement team, governance, security, and clinical informatics. Translate complex trade-offs into language non-architect audiences can engage with.

  • Establish and evolve the disaster recovery, backup, and reprocessing strategy across all platform layers. RPO under 5 minutes for streaming, RTO under 4 hours for full platform recovery.

Required Skills and Qualifications

  • Bachelor's or Master's degree in Computer Science, Engineering, or a related quantitative field

  • 8+ years of data engineering experience with at least 3-5 years in a Data Architect or Lead Data Engineer role

  • Deep expertise architecting and operating large-scale data platforms on Google Cloud Platform — Cloud Storage, Dataproc (specifically with Flink and Spark workloads), Cloud Composer, Secret Manager, Workload Identity, IAM, networking. BigQuery experience is useful background but is not used as the primary warehouse on this project

  • Hands-on experience with Apache Iceberg in production — table properties, partitioning strategies, snapshot semantics, schema evolution, CoW vs MoR write modes, compaction operations, and integration with metastore catalogs (BigLake, Glue, Polaris, or Nessie)

  • Production experience with Starburst (Galaxy or Enterprise) or Trino for analytical workloads against open table formats. Understanding of query planning, predicate pushdown, partition pruning, and workload separation patterns

  • Strong experience with streaming architectures using Apache Flink and Apache Kafka — stateful stream processing, exactly-once semantics, checkpointing, backpressure handling, and integration with Iceberg sinks

  • Expert SQL and strong programming skills in Python. Familiarity with PySpark and PyFlink for batch and streaming respectively

  • Deep experience with dbt — model materialization strategies, macros, tests, sources, exposures, project structure for large model graphs (hundreds of models). Specific experience with dbt-trino adapter is a strong plus

  • Comprehensive understanding of healthcare data standards: HL7v2 (ADT, ORU, ORM message types and segment-level parsing), CCDA, and FHIR R4 with US Core 6.1 profiles. Experience with FHIR Bundle assembly, profile validation, and resource versioning semantics

  • Hands-on experience with FHIR runtime platforms — FHIR-Repository or HAPI FHIR — including the interceptor framework, MDM module, and channel/subscription mechanisms

  • Experience with master data management for patient identity resolution. Informatica MDM specifically is preferred; equivalent experience with Verato, NextGate, or QuadraMed translates

  • Experience designing for HIPAA-regulated environments — PHI handling discipline, encryption at rest and in transit, audit logging conventions, BAA-relevant vendor decisions

  • Demonstrated ability to lead complex technical initiatives, make critical architectural decisions under uncertainty, and influence diverse stakeholders. Excellent written and verbal communication.

Nice-to-Have Skills

  • GCP Professional Data Engineer or Cloud Architect certification

  • Experience with Atlan or comparable governance platforms (Collibra, Alation, Unity Catalog)

  • Experience with multi-region active-active or active-passive deployments on GCP

  • Production experience operating Confluent Cloud at significant scale

  • Familiarity with spec-driven development workflows, particularly with AI-assisted code generation

  • Experience replacing or sunsetting legacy healthcare data platforms (Optum Health Data Engine, Innovaccer, Health Catalyst, Arcadia) is unique and highly relevant

If you like wild growth and working with happy, enthusiastic over-achievers, you'll enjoy your career with us!

Experience Level

Senior Level

Job role

Work location
Work locationIN MH Mumbai Eureka, India
Department
DepartmentSoftware Engineering
Role / Category
Role / CategoryDBA / Data warehousing
Employment type
Employment typeFull Time
Shift
ShiftDay Shift

Job requirements

Experience
ExperienceMin. 8 years

About company

Name
NameQuantiphi Analytics Solution Pvt Ltd
Job posted by Quantiphi Analytics Solution Pvt Ltd

Similar jobs you can apply for

Software / Web Developer

Web Developer

Zarna Enterprises
Mumbai/Bombay
₹3,000 - ₹5,000
Work from Office
Part Time
Full Time
Min. 5 years
Basic English
Omfurn India Limited

Engineering Trainee

Omfurn India Limited
Borivali East, Mumbai/Bombay
₹20,000 - ₹35,000
Work from Office
Full Time
Any experience
Basic English
Mcm Bpo Private Limited

AI Automation Specialist

Mcm Bpo Private Limited
Jogeshwari West, Mumbai/Bombay
₹10,000 - ₹57,000*
Work from Office
Full Time
Any experience
Basic English
Pioneer Refrigeration And Airconditioning Works/Dominion Oilfield Resources Rental

QA / QC Executive

Pioneer Refrigeration And Airconditioning Works/Dominion Oilfield Resources Rental
Byculla, Mumbai/Bombay
₹25,000 - ₹30,000
Work from Office
Full Time
Min. 2 years
Good (Intermediate / Advanced) English

Quality Control Engineer

Y J Associates
Mumbai/Bombay
₹50,000 - ₹90,000
Work from Office
Full Time
Min. 3 years
Basic English
R&D Security Solutions

Automation Engineer

R&D Security Solutions
Borivali West, Mumbai/Bombay
₹25,000 - ₹40,000*
Field Job
Full Time
Min. 2 years
Basic English

You can expect a minimum salary of 0 INR. The salary offered will depend on your skills, experience and performance in the interview.

The candidate should have completed the required education and people who have 8 to 13 years are eligible to apply for this job. You can apply for more jobs in Mumbai/Bombay to get hired quickly.

The candidate should have sound communication skills and sound communication skills for this job.

Both Male and Female candidates can apply for this job.

No, it's not a work from home job and can't be done online. You can explore and apply for other work from home jobs in Mumbai/Bombay at apna.

No work-related deposit needs to be made during your employment with the company.

Go to the apna app and apply for this job. Click on the apply button and call HR directly to schedule your interview.

The last date to apply for this job is . For more details, download apna app and find Full Time jobs in Mumbai/Bombay . Through apna, you can find jobs in 64 cities across India. Join NOW!