Digital : Google Data Engineering
Tata Consultancy Services
4 - 7 years
Bengaluru
Posted: 18/03/2026
Getting a referral is 5x more effective than applying directly
Job Description
TCS Hiring !!
About the Role:
Role: Digital : Google Data Engineering
Experience: 4-7 Years
Location: TCS - Bangalore||TCS - Hyderabad
Responsibilities
- Please read Job description before Applying
- NOTE: If the skills/profile matches and interested, please reply to this email by attaching your latest updated CV and with below few details:
Qualifications
- Name:
- Contact Number:
- Email ID:
- Highest Qualification in: (Eg. B.Tech/B.E./M.Tech/MCA/M.Sc./MS/BCA/B.Sc./Etc.)
- Current Organization Name:
- Total IT Experience:
- Location: Chennai/Kolkata/Hyderabad/Pune/Bangalore/Nagpur
- Current CTC:
- Expected CTC:
- Notice period:
- Whether worked with TCS - Y/N
- Experience with dashboarding tools like plx dashboard and looker studio Experience with building data pipelines, reports, best practices and frameworks.
- Experience with design and development of scalable and actionable solutions (dashboards, automated collateral, web applications).
- Experience with code refactoring for optimal performance.
- Experience writing and maintaining ETLs which operate on a variety of structured and unstructured sources.
- Familiarity with non-relational data storage systems (NoSQL and distributed database management systems).
Skills
- Strong proficiency in SQL, NoSQL, ETL tools, BigQuery and at least one programming language (e.g., Python, Java).
- Big Query,Data Flow,Data Proc,Cloud Sql,Teraform etc
- Strong understanding of data structures, algorithms, and software design principles.
- Experience with data modeling techniques and methodologies.
- Proficiency in troubleshooting and debugging complex data-related issues.
- Ability to work independently and as part of a team.
Responsibilities
- Data Pipeline Development: Design, implement, and maintain robust and scalable data pipelines to extract, transform, and load data from various sources into our data warehouse or data lake.
- Data Modeling and Warehousing: Collaborate with data scientists and analysts to design and implement data models that optimize query performance and support complex analytical workloads.
- Cloud Infrastructure: Leverage Google Cloud and other internal storage platforms to build and manage scalable and cost-effective data storage and processing solutions.
- Data Quality Assurance: Implement data quality checks and monitoring processes to ensure the accuracy, completeness, and consistency of data.
- Build large scale data and analytics solutions on GCP, Efficiently use the GCP platform to integrate large datasets from multiple data sources, analyse data, data modelling, data exploitation/visualization, DevOps, CI/CD implementation Build automated data pipelines and work in Data engineering solution on GCP using Cloud BigQuery, Cloud DataProc,
- Performance Optimization: Continuously monitor and optimize data pipelines and queries for performance and efficiency.
- Collaboration: Work closely with data scientists, analysts, and other stakeholders to understand their data needs and deliver solutions that meet their requirements.
Desirable
- Experience Cloud Storage or equivalent cloud platforms
- Knowledge of BigQuery ingress and egress patterns
- Experience in writing Airflow DAGs
- Knowledge of pubsub,dataflow or any declarative data pipeline tools using batch and streaming ingestion
- Other GCP Services: Vertex AI
Services you might be interested in
Improve Your Resume Today
Boost your chances with professional resume services!
Get expert-reviewed, ATS-optimized resumes tailored for your experience level. Start your journey now.
