Job Summary:
We are looking for an experienced Data Architect to design and manage scalable data systems, with a focus on real-time data processing and integration. This role requires proficiency in event-driven architecture, data streaming via Kafka, and API-based data exposure to support dynamic, data-driven decision-making across the organization. The ideal candidate will have a strong background in data engineering, cloud technologies, and data governance, as well as experience with both batch and streaming data pipelines.
Responsibilities:
Data Architecture and Real-Time Integration
- Design and implement data architectures to support real-time, event-driven data models.
- Architect and optimize data flows using event streaming platforms like Apache Kafka.
Event-Driven Architecture
- Develop and maintain event-driven architectures that support key business functions, utilizing pub/sub and other messaging patterns.
- Define data streams and topics in Kafka and ensure efficient and scalable processing of real-time data.
API Development and Data Exposure
- Design and manage APIs to expose data to downstream applications and systems.
- Collaborate with development teams to build RESTful APIs endpoints for real-time data access.
Cloud-Based Data Solutions
- Implement and optimize data architectures in cloud environments (e.g., AWS, Azure, or Google Cloud Platform).
- Leverage cloud-native tools to build scalable data pipelines and storage solutions.
Data Governance and Quality
- Ensure data governance frameworks, including data quality, security, and compliance, are integrated across real-time and batch processes.
- Monitor and enhance data quality, consistency, and accuracy across all systems.
Collaboration and Stakeholder Management
- Work closely with data engineering, analytics, and business teams to gather requirements and design data solutions that support business needs.
- Provide mentorship to junior data engineers and architects and serve as a thought leader in data architecture best practices.
Qualifications:
- Bachelor’s or Master’s degree in Computer Science, Data Science, Information Systems, or a related field.
- 10+ years of experience in data architecture, data modeling, and database design.
- Proficiency in event-driven architectures and tools like Apache Kafka.
- Strong experience with API design and integration to expose data to downstream systems.
- Expertise in cloud data services (AWS, Azure, or GCP) and distributed data systems.
- Knowledge of data modeling techniques for both batch and streaming data.
- Familiarity with data governance and security frameworks.