Posted on: 24/01/2026
Description :
About the job
Key Responsibilities :
- Build and optimize ETL/ELT pipelines using Python and PySpark
- Work with structured and semi-structured data using SQL
- Ingest data from multiple sources (APIs, databases, streaming, files)
- Perform data cleaning, transformation, and validation
- Implement data quality checks and monitoring
- Optimize data processing for performance and cost
- Collaborate with data scientists, analysts, and stakeholders
- Support the entire data engineering lifecycle : ingestion processing storage analytics
Mandatory Technical Skills :
- Programming : Python, PySpark
- Databases : Strong SQL (PostgreSQL / MySQL / Redshift)
- Data Processing : Batch and streaming data pipelines
- ETL Tools : AWS Glue / Spark-based frameworks
Good to Have :
- Experience with Airflow or AWS Step Functions
- Knowledge of Delta Lake / Iceberg / Hudi
- Exposure to CI/CD for data pipelines
- Experience with data warehousing and analytics
Education : B.E / B.Tech / M.Tech / MCA or equivalent
Did you find something suspicious?
Posted by
Posted in
Data Engineering
Functional Area
Data Engineering
Job Code
1605906