Pyspark Developer (Pan India)
Tata Consultancy Services
2 - 5 years
Bengaluru
Posted: 08/01/2026
Getting a referral is 5x more effective than applying directly
Job Description
Skill:Pyspark Developer
Experience range :5+
Location :Pan India
Must-Have
- 5+ years of experience in data engineering, with strong focus on PySpark/Spark for big data processing.
- Expertise in building data pipelines and ingestion frameworks from relational, semi-structured (JSON, XML), and unstructured sources (logs, PDFs).
- Proficiency in Python with strong knowledge of data processing libraries.
- Strong SQL skills for querying and validating data in platforms like Amazon Redshift, PostgreSQL, or similar.
- Experience with distributed computing frameworks (e.g., Spark on EMR, Databricks).
- Familiarity with workflow orchestration tools (e.g., AWS Step Functions, or similar).
- Solid understanding of data lake / data warehouse architectures and data modeling basics.
1. Experience with AWS data services: Glue, S3, Redshift, Lambda, CloudWatch, etc.
2. Familiarity with Delta Lake or similar for large-scale data storage.
3 Exposure to real-time streaming frameworks (e.g., Spark Structured Streaming, Kafka).
4 Knowledge of data governance, lineage, and cataloging tools (e.g., AWS Glue Catalog, Apache Atlas).
5 Understanding of DevOps/CI-CD pipelines for data projects using Git, Jenkins, or similar tools.
Services you might be interested in
Improve Your Resume Today
Boost your chances with professional resume services!
Get expert-reviewed, ATS-optimized resumes tailored for your experience level. Start your journey now.
