Posted on: 25/08/2025
Kafka Administration & Management :
- Experience with and Understanding of cloud based messaging services like AWS MSK, Confluent Cloud, Azure Event Hubs, or GCP Pub/Sub).
- Configure and tune Kafka brokers, ZooKeeper, topics, partitions, and replication to ensure optimal performance and scalability.
- Implement Kafka security best practices, including SSL/TLS encryption, Kerberos authentication, ACLs, and role-based access control (RBAC).
- Monitor Kafka clusters using tools like Prometheus, Grafana, Confluent Control Center, and OpenTelemetry to track system health, resource utilization, and performance metrics.
- Troubleshoot Kafka broker failures, consumer/producer lag, replication issues, and other system bottlenecks.
- Manage Kafka Schema Registry, Kafka Connect, and Kafka Streams for seamless data integration across various platforms.
- Implement disaster recovery strategies, including cross-cluster replication with MirrorMaker 2.0.
Real-Time & Batch Data Processing (Flink/Spark) :
- Implement complex event processing (CEP) and windowed aggregations in Flink for business analytics.
- Build and optimize Apache Spark jobs for batch data processing, ensuring high performance and cost efficiency.
- Integrate Flink/Spark applications with Kafka, Hadoop, S3, Snowflake, and NoSQL databases.
- Tune Flink and Spark performance parameters, including checkpointing, parallelism, and memory management.
Data Infrastructure & Governance:
- Implement observability using distributed tracing, logging, and monitoring tools like Jaeger, ELK Stack, and Splunk.
- Collaborate with DevOps teams to automate Kafka infrastructure using Terraform, Ansible, or CloudFormation.
- Ensure data governance, lineage, and compliance (GDPR, HIPAA, SOC 2) by integrating Apache Atlas, Confluent RBAC, or Ranger.
- Work with data teams, ML engineers, and business stakeholders to build scalable and efficient data solutions.
Required Qualifications :
- Expert-level knowledge of Apache Kafka internals, including broker configuration, producer/consumer tuning, and security.
- Strong hands-on experience in Kafka Streams, Kafka Connect, MirrorMaker 2.0, and Schema Registry.
- Proficiency in Apache Flink (or Spark Streaming) for stream processing.
- Strong programming skills in Python, Java, or Scala.
- Experience with SQL and NoSQL databases (e.g., PostgreSQL, Cassandra, MongoDB).
- Hands-on experience with cloud platforms (AWS, Azure, GCP) and Kubernetes.
- Knowledge of monitoring tools (Prometheus, Grafana, Confluent Control Center, ELK).
Preferred Qualifications :
- Hands-on knowledge of Kafka Tiered Storage and Cloud-native Kafka deployments.
- Knowledge of workflow orchestration tools like Apache Airflow, Prefect, or Dagster.
- Contributions to open-source Kafka/Flink/Spark projects
Did you find something suspicious?
Posted By
Posted in
DevOps / SRE
Functional Area
DevOps / Cloud
Job Code
1535667
Interview Questions for you
View All