Posted on: 25/07/2025
Thryve Digital Health LLP is a fast-growing global healthcare partner delivering strategic innovation and IT expertise to healthcare organizations.
As a captive of a leading US-based healthcare conglomerate, Thryve brings unique insights into healthcare operations, delivering next-generation solutions that help our partners deliver superior consumer experiences.
Our commitment to integrity, diversity, and inclusion strengthens our position as an equal opportunity employer.
Role Summary :
We are looking for a Lead Data Engineer with strong expertise in Apache Kafka, Apache Flink, and real-time data processing.
This role will be responsible for building and optimizing data pipelines, stream processing applications, and scalable ingestion architectures using Confluent Kafka ecosystem components.
If you're passionate about building high-performance, real-time data systems in a fast-paced collaborative environment, this is the role for you.
Key Responsibilities :
- Design and develop scalable real-time and batch data pipelines using Kafka, Flink, Kafka Connect, and ksqlDB.
- Implement and manage Kafka Connectors to integrate with diverse data sources (APIs, databases, message queues).
- Build robust Flink applications for complex event processing, enrichment, and real-time analytics.
- Write and optimize ksqlDB queries for transformations, filters, and aggregations.
- Implement data quality monitoring, validation, and automated testing for pipelines.
- Identify performance bottlenecks and optimize streaming architectures.
- Contribute to DevOps automation for deployment, monitoring, and maintenance of data pipelines.
- Stay updated with industry best practices in streaming technologies and lead internal adoption of standards.
- Collaborate with architects and tech leads in India and the US to deliver Proof of Concepts (POCs) and Minimum Viable Products (MVPs).
- Participate in Agile ceremonies, code reviews, and provide project updates to managers.
Required Qualifications :
- Bachelors or Masters degree in Computer Science, Engineering, or related discipline.
- 8+ years of experience in Data Engineering, with a strong background in Kafka and Big Data systems.
- Deep expertise in Apache Flink for stream processing.
- Strong experience with Confluent Kafka, including Kafka Brokers, Producers/Consumers, and Schema Registry.
- Solid knowledge of Kafka Connect and ksqlDB.
- Proficient in building and deploying large-scale ETL/ELT and ingestion frameworks.
- Hands-on experience with big data stack on any cloud platform (GCP, AWS, or Azure).
- Self-motivated with the ability to work both independently and within a team.
Preferred Qualifications :
- Experience working on Google Cloud Platform (GCP).
- Prior exposure to the Healthcare domain and data compliance practices.
- Familiarity with Agile/Scrum development methodologies
Did you find something suspicious?
Posted By
Posted in
Data Engineering
Functional Area
Data Engineering
Job Code
1519475
Interview Questions for you
View All