HamburgerMenu
hirist

Job Description

Job Title : GCP Data Architect

Location : Bangalore

Job Summary :

We are seeking a highly skilled GCP Data Architect with extensive experience in architecting and implementing large-scale, cloud-native data solutions. The ideal candidate will have strong expertise in Google Cloud Platform (GCP), data modeling, data governance, ETL/ELT pipelines, and advanced analytics ecosystems. This role involves designing enterprise-grade data platforms that power business intelligence, AI/ML, and real-time decision-making across structured, semi-structured, and unstructured datasets.

Key Responsibilities :

Enterprise Data Architecture & Modeling :

- Design end-to-end data architecture frameworks on GCP to support enterprise data strategy.

- Develop logical, physical, and canonical data models supporting OLTP, OLAP, and streaming workloads.

- Ensure scalability, security, and performance alignment with business use cases.

Data Integration & Pipelines :

- Architect and implement ETL/ELT pipelines using Cloud Dataflow, Apache Beam, Dataproc, and Composer (Airflow).

- Enable real-time and batch data ingestion using Pub/Sub, Kafka, or equivalent.

- Optimize data pipeline orchestration and reliability for mission-critical workloads.

Data Warehousing & Analytics :

- Design, implement, and optimize data lakes and warehouses leveraging BigQuery, Cloud Storage, and Dataform.

- Develop partitioning, clustering, and materialized views to improve query performance and cost efficiency.

- Integrate GCP solutions with BI tools (e.g., Looker, Tableau, Power BI).

Cloud Migration & Modernization :

- Lead migration of legacy data platforms to GCP, ensuring minimal downtime and high availability.

- Re-engineer monolithic data systems into cloud-native, microservices-driven architectures.

Performance & Optimization :

- Perform SQL query tuning, indexing strategies, and workload management for high-performance data solutions.

- Conduct capacity planning, monitoring, and auto-scaling to ensure SLA compliance.

Data Governance, Security & Compliance :

- Implement data governance frameworks, metadata management, and data catalogs (Data Catalog, Collibra, or equivalent).

- Ensure compliance with GDPR, HIPAA, SOC 2, or RBI guidelines for financial data.

- Enforce IAM roles, VPC Service Controls, encryption (KMS), and DLP policies for secure data handling.

Collaboration & Leadership :

- Partner with business stakeholders, product teams, and data scientists to translate analytical needs into scalable architectures.

- Provide technical leadership, mentorship, and best practices to data engineers and developers.

Qualifications :

Education : Bachelors or Masters degree in Computer Science, Information Technology, Data Engineering, or related field.

Experience : 8+ years in data architecture, with at least 3+ years on GCP ecosystem.

Proven track record in :

- Big Data technologies (Hadoop, Spark, Kafka, Flink).

- Cloud-native services (BigQuery, Dataflow, Dataproc, Pub/Sub, Composer).

- Databases : Proficiency in both SQL (PostgreSQL, MySQL, Oracle) and NoSQL (MongoDB, Cassandra, Firestore).

- ETL/ELT tools : Informatica, Talend, dbt, or equivalent.

- Strong expertise in data modeling, data warehousing, and data migration strategies.

- Hands-on with DevOps practices CI/CD pipelines (Cloud Build, Jenkins, GitLab), Infrastructure as Code (Terraform, Deployment Manager).

- Excellent problem-solving, stakeholder management, and communication skills.

Preferred Skills & Certifications :

Certifications :

- Google Cloud Professional Data Engineer / Cloud Architect.

- AWS or Azure certifications (multi-cloud exposure preferred).

- Knowledge of machine learning pipelines (Vertex AI, TensorFlow Extended).

- Familiarity with containerization and orchestration (Docker, Kubernetes, GKE).

- Experience with data governance tools like Collibra, Alation, or Atlan.


info-icon

Did you find something suspicious?