Posted on: 02/02/2026
Description :
Experience with Elasticsearch, streaming data, and modern analytics platforms is preferred.
Mandatory Skills Requirements :
- Extensive experience with Apache Spark Structured Streaming for near real-time and streaming data processing.
- Strong hands-on experience with Apache Kafka, including integration with Spark for reliable real-time data ingestion and event-driven pipelines.
- Experience working with analytical and distributed data stores such as ClickHouse, Trino/Presto, and data lake technologies (Delta Lake or equivalent).
- Solid understanding of data modeling and metric design for large-scale analytics systems, including fact/dimension modeling and event-based schemas.
- Proven ability to design and implement ETL / ELT pipelines for data ingestion, transformation, aggregation, and performance optimization using Spark.
- Demonstrated experience in writing efficient, scalable, and maintainable code for large-scale data processing workloads.
- Experience operating in on-prem or hybrid data platforms, with a working understanding of cluster resource management, performance tuning, and capacity planning.
- Familiarity with Elasticsearch for search, observability, or analytical use cases is a plus.
Preferred Skills Requirements :
- Strong familiarity with version control systems, particularly Git, and collaborative development workflows.
- Working knowledge of cloud platforms such as AWS, Azure, or Google Cloud, primarily for data services, storage, or hybrid deployments.
- Understanding of distributed data systems and database administration principles, including performance tuning, reliability, and scaling of analytical or NoSQL databases (e.g., ClickHouse, Elasticsearch, HBase, or similar
Did you find something suspicious?
Posted by
Posted in
Data Engineering
Functional Area
Data Engineering
Job Code
1608842