Posted on: 05/11/2025
Description :
Responsibilities :
- Design, implement, and manage our AWS infrastructure, with a strong emphasis on automation, resilience, and cost efficiency.
- Implement and manage stream processing frameworks (Kafka).
- Handle orchestration and ETL workloads, employing services like AWS Glue, Athena, Redshift, or Apache Airflow.
- Develop P0 and P1 issues alert and resolution process/pipeline.
- Monitor, debug, and resolve production issues related to data and infrastructure in real time.
- Implement IAM controls, logging, alerts, and Security Best Practices across all components.
- Data migration and storage, monitor storage usage, and ensure data integrity and data security.
- Provide deployment automation (Docker) and collaborate with application engineers to enable smooth delivery.
Requirements :
- 2+ years of experience with AWS services (VPC, EC2 S3 Security Groups, RDS, MSK).
- Ability to handle 5 billion events/day and 1M+ concurrent users' workloads gracefully.
- Familiar with scripting (Python, Terraform, bash) and automation practices (Infrastructure as Code).
- Familiar with network fundamentals, Linux, scaling strategies, backup routines, and CDC pipeline and logging.
- Collaborative team player able to work with engineers, data analysts, and stakeholders.
Preferred Tools and Technologies :
- AWS : EC2 S3 VPC, Security Groups, RDS, DocumentDB, MSK, Glue, Athena, CloudWatch.
- Infrastructure as Code : Terraform.
- Scripted automation : Python, Bash.
- Container orchestration : Docker, ECS, or EKS.
- Workflow orchestration : Apache Airflow.
- Streaming framework : Apache Kafka.
- Database - Postgresql (must), Redis (must), Cassandra, MongoDB.
- Other : Linux, Git, Security best practices (IAM, Security Groups, ACM), jenkins.
Did you find something suspicious?
Posted By
Posted in
DevOps / SRE
Functional Area
DevOps / Cloud
Job Code
1570226
Interview Questions for you
View All