GCP Data Architect
TerraGiG
8 - 10 years
Bengaluru
Posted: 23/12/2025
Job Description
Role: GCP Data Architect
Duration: Permanent
Location: Chennai, Bangalore, Hyderabad, Pune, Vadodara
Timings: Full Time (As per company timings)
Notice Period: (Immediate Joiner - Only)
Experience: 8-10 Years
Key Responsibilities
Architecture & Design
Architect highly scalable, reliable, and secure cloud-native applications on GCP.
Define end-to-end solution architectures involving microservices, event-driven patterns, API ecosystems, and data pipelines.
Design MongoDB data models, indexing strategies, partitioning/sharding approaches, and performance optimization.
Architect Kafka-based streaming solutions including topic design, consumer group strategy, message serialization, and delivery semantics.
Ensure solutions follow best practices in scalability, resiliency, observability, security, and cost optimization.
Create architecture documents, sequence diagrams, logical/physical data models, and integration patterns.
Hands-on Development
Develop and review code in languages such as Python, Java, Node.js, or Go.
Build microservices, APIs, data ingestion jobs, Kafka producers/consumers, and GCP-native applications.
Perform POCs and build reusable frameworks, utilities, and accelerators.
Support CI/CD pipeline development and automated test practices.
Conduct code reviews, troubleshoot performance issues, and mentor developers.
Cloud Platform Expertise (GCP)
Strong knowledge of compute services (GKE, Cloud Run, GCE), storage (GCS), networking (VPC, Cloud Load Balancing), and IAM.
Experience with GCP data services such as BigQuery, Pub/Sub, Dataflow, Cloud Composer, Cloud Functions.
Optimize cloud workloads for performance and cost.
Design and implement observability Cloud Logging, Monitoring, APM.
MongoDB Expertise
Schema design for document databases aligned with access patterns and performance goals.
Experience with MongoDB Atlas, replica sets, sharding, backup/restore, and cluster tuning.
Strong understanding of query optimization, aggregations, and schema evolution.
Kafka Expertise
Experience with Kafka clusters (self-managed or Confluent Cloud).
Design and implement Kafka producers, consumers, stream processors.
Experience with Schema Registry, Connectors (Kafka Connect), KStreams/KSQL is a plus.
Tuning throughput, consumer lag handling, partition strategies, and ensuring high availability.
Cross-functional Collaboration
Work with product managers, business analysts, developers, and QA engineers to translate requirements into technical solutions.
Provide technical leadership, mentoring, and architectural governance.
Drive architecture reviews, performance reviews, and security reviews.
DevOps & Security
Experience in CI/CD tools (GitHub Actions, Jenkins, GitLab).
Knowledge of Terraform/Infrastructure-as-Code on GCP.
Implement secure coding practices, encryption, IAM, VPC design, secret management.
Perform threat modeling and ensure compliance with organizational standards.
Required Skills & Qualifications
8+ years of experience in software engineering with at least 4 years in solution architecture.
Strong hands-on expertise in GCP, MongoDB, and Kafka.
Proficient in one or more programming languages (Java, Python, Go, Node.js).
Understanding of microservices, distributed systems, caching, messaging, and API best practices.
Experience in designing high-throughput, low-latency systems.
Familiarity with containerization (Docker) and orchestration (Kubernetes).
Excellent analytical, technical documentation, and communication skills.
Interested candidates please share your resume to
Services you might be interested in
Improve Your Resume Today
Boost your chances with professional resume services!
Get expert-reviewed, ATS-optimized resumes tailored for your experience level. Start your journey now.
