Posted on: 09/10/2025
Description :
- Develop scalable data ingestion, transformation, and processing frameworks for high-volume data environments.
- Collaborate with product, analytics, and infrastructure teams to ensure data reliability, availability, and performance.
- Build APIs and data models that support large-scale analytics and machine learning
applications.
- Optimize query performance, storage systems, and ETL workflows for massive datasets.
- Mentor junior engineers, contribute to best practices, and lead design reviews for critical data
infrastructure.
Requirements :
- 4+ years of experience in data engineering, backend development, or large-scale data
systems.
- Strong proficiency in Python; experience with Java or Scala is a plus.
- Hands-on experience with real-time processing frameworks such as Kafka, Spark Streaming,
or Flink.
- Solid understanding of data modeling, distributed systems, and API integrations.
- Proficiency with SQL and NoSQL databases, and working experience on AWS (S3 EMR,
Redshift, Glue, Kinesis).
- Familiarity with web scraping frameworks (Scrapy, Selenium, Playwright) or similar data acquisition tools.
- Experience in optimizing ETL pipelines and handling large-scale datasets.
Preferred Qualifications :
- Experience in AI/ML data pipelines or analytical dashboarding.
- Exposure to workflow orchestration tools (Airflow, Step Functions, or similar).
- Experience mentoring engineers or leading small technical teams.
Did you find something suspicious?
Posted by
Posted in
Data Engineering
Functional Area
Data Engineering
Job Code
1557562