GCP Big Data Developer
Impetus
2 - 5 years
Bengaluru
Posted: 10/12/2025
Getting a referral is 5x more effective than applying directly
Job Description
```html
About the Company
Must have: Bigdata, GCP, Dataproc, Dataflow, SQL, Spark
About the Role
Must have: Operating knowledge of cloud computing platforms (GCP, especially Big Query, Dataflow, Dataproc, Storage, VMs, Networking, Pub Sub, Cloud Functions, Composer services).
Responsibilities
- Should be aware with columnar database e.g. parquet, ORC etc.
- Visualize and evangelize next generation infrastructure in Cloud platform/Big Data space (Batch, Near Real-time, Real-time technologies).
- Passionate for continuous learning, experimenting, applying and contributing towards cutting edge open source technologies and software paradigms.
- Developing and implementing an overall organizational data strategy that is in line with business processes.
- The strategy includes data model designs, database development standards, implementation and management of data warehouses and data analytics systems.
Qualifications
- Expert-level proficiency in at least 4-5 GCP services.
Required Skills
- Experience with technical solutions based on industry standards using GCP - IaaS, PaaS and SaaS capabilities.
- Strong understanding and experience in distributed computing frameworks.
- Experience working within a Linux computing environment, and use of command line tools including knowledge of shell/Python scripting for automating common tasks.
Preferred Skills
- None specified.
Pay range and compensation package
None specified.
Equal Opportunity Statement
None specified.
```
Services you might be interested in
Improve Your Resume Today
Boost your chances with professional resume services!
Get expert-reviewed, ATS-optimized resumes tailored for your experience level. Start your journey now.
