Pyspark developer
Tata Consultancy Services
2 - 5 years
Chennai
Posted: 08/01/2026
Getting a referral is 5x more effective than applying directly
Job Description
Role - Pyspark developer
Required Technical Skill Set -Pyspark, Redshift, PostgreSQL
year of experience- 4 to 8 years
Location- Chennai / Hyderabad
Desired Competencies (Technical/Behavioral Competency)
Must-Have
- 5+ years of experience in data engineering, with strong focus on PySpark/Spark for big data processing.
- Expertise in building data pipelines and ingestion frameworks from relational, semi-structured (JSON, XML), and unstructured sources (logs, PDFs).
- Proficiency in Python with strong knowledge of data processing libraries.
- Strong SQL skills for querying and validating data in platforms like Amazon Redshift, PostgreSQL, or similar.
- Experience with distributed computing frameworks (e.g., Spark on EMR, Databricks).
- Familiarity with workflow orchestration tools (e.g., AWS Step Functions, or similar).
- Solid understanding of data lake / data warehouse architectures and data modeling basics.
Good-to-Have
- Experience with AWS data services: Glue, S3, Redshift, Lambda, CloudWatch, etc.
- Familiarity with Delta Lake or similar for large-scale data storage.
- Exposure to real-time streaming frameworks (e.g., Spark Structured Streaming, Kafka).
- Knowledge of data governance, lineage, and cataloging tools (e.g., AWS Glue Catalog, Apache Atlas).
- Understanding of DevOps/CI-CD pipelines for data projects using Git, Jenkins, or similar tools.
Services you might be interested in
Improve Your Resume Today
Boost your chances with professional resume services!
Get expert-reviewed, ATS-optimized resumes tailored for your experience level. Start your journey now.
