Posted on: 08/10/2025
Description :
- Develop scalable data ingestion, transformation, and processing frameworks for high-volume data environments.
- Collaborate with product, analytics, and infrastructure teams to ensure data reliability, availability, and performance.
- Build APIs and data models that support large-scale analytics and machine learning
applications.
- Optimize query performance, storage systems, and ETL workflows for massive datasets.
- Mentor junior engineers, contribute to best practices, and lead design reviews for critical data
infrastructure.
Requirements :
- 4+ years of experience in data engineering, backend development, or large-scale data
systems.
- Strong proficiency in Python; experience with Java or Scala is a plus.
- Hands-on experience with real-time processing frameworks such as Kafka, Spark Streaming,
or Flink.
- Solid understanding of data modeling, distributed systems, and API integrations.
- Proficiency with SQL and NoSQL databases, and working experience on AWS (S3 EMR,
Redshift, Glue, Kinesis).
- Familiarity with web scraping frameworks (Scrapy, Selenium, Playwright) or similar data acquisition tools.
- Experience in optimizing ETL pipelines and handling large-scale datasets.
Preferred Qualifications :
- Experience in AI/ML data pipelines or analytical dashboarding.
- Exposure to workflow orchestration tools (Airflow, Step Functions, or similar).
- Experience mentoring engineers or leading small technical teams.
Did you find something suspicious?
Posted By
Posted in
Data Engineering
Functional Area
Data Engineering
Job Code
1557562
Interview Questions for you
View All