- Company Name
- BlueRose Technologies
- Job Title
- Data Architect
- Job Description
-
**Job title**
Data Architect
**Role Summary**
Design and govern end‑to‑end data product architecture across streaming, batch, and lakehouse environments, ensuring scalability, reliability, and compliance with data‑mesh principles.
**Expectations**
* Deliver production‑ready data pipelines and architectures in a contract, hybrid setting.
* Own data quality, governance, and cost monitoring for large‑scale event streams and analytics workloads.
**Key Responsibilities**
* Architect data products using Kafka (Confluent), AWS MSK, Kinesis, and EventBridge; enforce ordering, replay, and semantic guarantees.
* Manage schema evolution with Avro/Protobuf in a Schema Registry; define subject strategy and compatibility rules.
* Build streaming pipelines: Kinesis → S3 → Glue (batch and streaming) → Athena/Redshift; implement DLQ and back‑pressure patterns.
* Apply lakehouse strategies (Iceberg) and partitioning for performance and cost efficiency.
* Design payment processing flows (ISO 20022 PAIN/PACS/CAMT) and reconcilation logic for APIs, files, and SWIFT.
* Create observability dashboards, alerts, and FinOps KPIs; troubleshoot latency and throughput at scale.
* Produce high‑quality code, conduct reviews, automate tests, and embed security best practices.
* Model logical data with 1NF/BCNF and physical schemas for OLTP/OLAP; evaluate denormalization, star, and Data Vault designs.
* Implement CQRS/ES patterns where appropriate; enforce domain‐driven design, bounded contexts, and event contracts.
**Required Skills**
* Proven data product architecture experience.
* Proficiency with Kafka (Confluent), AWS MSK, Kinesis, EventBridge.
* Schema registry expertise (Avro/Protobuf) and event semantics.
* AWS data stack: S3, Glue, Athena, Redshift, Step Functions, Lambda, Iceberg.
* Payment domain knowledge: ISO 20022, PAIN/PACS/CAMT, reconciliation.
* Data‑mesh mindset: ownership, SLAs, lineage, retention.
* Observability (Dashboards, alerts) and FinOps (cost KPIs).
* Strong coding in Python/Java/Scala; automated testing, CI/CD pipelines.
* Logical/physical data modeling (ER, normalization, denormalization, SCD).
* CQRS, Event‑Sourcing, Saga, and domain‑driven design fundamentals.
* Experience with QuickSight/Tableau (nice to have).
**Required Education & Certifications**
* Bachelor’s (or higher) in Computer Science, Data Engineering, or related field.
* Relevant certifications (e.g., AWS Certified Solutions Architect, Confluent Certified Developer) preferred.