Data Engineer - GCP stack ( Mumbai/pune/Bangalore)
AgileEngine
8 - 10 years
Pune City
Posted: 15/05/2026
Getting a referral is 5x more effective than applying directly
Job Description
Immediate Hiring | Fast-Track Interviews | Work on Global Data Platforms
Were hiring a Data Engineer (GCP) to join a high-impact data platform initiative at . If you enjoy building scalable data pipelines, working with financial datasets, and leveraging modern cloud + AI technologies, this role is for you.
Location: Mumbai / Pune / Bangalore - 2 Days Work from Office.
---
Why You Should Apply:
Work on One MSCI Data Platform (OMDP) a global data transformation initiative
Build large-scale, regulatory-grade financial data systems
Exposure to AI-driven data platforms (Vertex AI, anomaly detection, semantic search)
Collaborate with global engineering, data, and compliance teams
Strong ownership with real impact on business-critical systems
---
What We Need (Non-Negotiable):
68 years of experience in Data Engineering
Strong Python skills (data pipelines, transformation, automation)
Advanced SQL + BigQuery (partitioning, clustering, materialized views, time-series queries)
Hands-on experience with Cloud Composer (Apache Airflow) DAGs, SLA, retries
Experience with Dataproc (Apache Spark) batch ingestion & Delta Lake
Strong knowledge of GCP ecosystem (Cloud Storage, Pub/Sub, Datastream, IAM, Monitoring)
Experience with Git workflows (branching, PRs, pipeline-as-code)
Familiarity with REST APIs integrations
Experience using AI-assisted dev tools (Copilot, Cursor, etc.)
---
What Youll Do:
Build scalable, fault-tolerant data pipelines on GCP
Design BigQuery lakehouse layers & Dataproc-driven workflows
Implement bitemporal data models (valid-time + system-time)
Develop and maintain data quality & testing frameworks
Own end-to-end pipeline lifecycle ingestion QA audit delivery
Work with global stakeholders across engineering, business & compliance
Support AI integrations (data quality, anomaly detection, semantic search using Vertex AI)
Contribute to shared data platform services (not siloed development)
---
Bonus (Good to Have):
Experience with pandas / PySpark
Knowledge of Dataplex, ClickHouse, CDC (Datastream)
Understanding of financial data (equities, fixed income, corporate actions)
Experience with BigQuery cost optimization (slots, reservations)
Exposure to Terraform & CI/CD pipelines
Experience with LLMs / Agentic AI use cases
---
Priority Given To Candidates Who:
Have strong hands-on GCP + Airflow experience
Have worked on large-scale data platforms / financial datasets
Can join within 2 weeks
Demonstrate strong ownership and problem-solving mindset
---
Serious applicants only. Profiles without GCP + Airflow hands-on experience will not be considered.
Pls read below critical requirements for this position. We can proceed only if you have for all the mentioned criterias
Do you have 68 years of hands-on Data Engineering experience building production-grade pipelines?
Do you have strong Python expertise for data pipelines, transformations, and automation (not just scripting)?
Do you have deep SQL + BigQuery experience (partitioning, clustering, materialized views, large-scale queries)?
Do you have real hands-on experience with Cloud Composer / Apache Airflow (writing DAGs, managing dependencies, SLAs, retries)?
Do you have practical experience with Dataproc / Apache Spark (batch ingestion, incremental loads, Delta Lake merges)?
Do you have strong exposure to GCP ecosystem (Cloud Storage, Pub/Sub, Datastream, IAM, Monitoring)?
Do you actively use Git for collaboration (branching, PR reviews, CI/CD, pipeline-as-code)?
Do you have experience integrating REST APIs for external data ingestion?
Do you use AI-assisted coding tools (Copilot, Cursor, etc.) to improve productivity?
Do you have experience building scalable, fault-tolerant data pipelines handling large datasets?
Do you understand or have worked with time-series or bitemporal data models (huge plus)?
Are you based in (or open to relocating to) Mumbai / Pune / Bangalore with 2 days WFO?
Are you comfortable with a role that may start as a 4-month client engagement, but is a permanent position with AgileEngine and can extend or transition to other long-term/WFH projects?
Can you join within 2 weeks or less?
. To Apply Share:
1. Email ID
2. years of Experience
3. Current CTC / Expected CTC
4. Notice Period
5. Location
Services you might be interested in
Improve Your Resume Today
Boost your chances with professional resume services!
Get expert-reviewed, ATS-optimized resumes tailored for your experience level. Start your journey now.
