HamburgerMenu
hirist

Big Data Engineer - Hadoop/Spark

HIRE YOO INFOTECH PRIVATE LIMITED
Multiple Locations
3 - 5 Years

Posted on: 13/07/2025

Job Description

Role Overview :


As a Big Data Engineer, you will be responsible for building robust data pipelines, developing scalable data processing systems, and optimizing data workflows across distributed environments.

You will collaborate closely with data scientists, analysts, and software engineers to ensure efficient data flow and accessibility across platforms.


Key Responsibilities :


- Design, develop, and manage scalable data pipelines using big data technologies.


- Build data integration workflows for structured and unstructured data sources.

- Implement ETL/ELT processes for real-time and batch processing.

- Optimize data processing jobs for performance, scalability, and fault tolerance.

- Ensure data quality, security, and governance throughout the data lifecycle.

- Collaborate with analytics and engineering teams to support data product development.

- Monitor and maintain large-scale distributed systems and troubleshoot performance issues.

- Maintain documentation for data architecture, pipeline configurations, and operational procedures.


Required Skills & Qualifications :

- 3+ years of hands-on experience in Big Data engineering.

- Proficiency in technologies such as Hadoop, Spark, Hive, Kafka, Flink, or Presto.

- Strong programming/scripting skills in Python, Java, or Scala.

- Experience with cloud-based data platforms (AWS EMR, GCP BigQuery, Azure Data Lake).

- Familiarity with data warehousing and distributed computing systems.

- Solid understanding of SQL, data modeling, and query optimization.

- Experience with workflow orchestration tools like Airflow, Luigi, or Dagster.

- Knowledge of version control (Git) and CI/CD pipelines for data deployments.


Preferred Skills :


- Experience with real-time streaming data pipelines using Kafka, Flink, or Kinesis.


- Familiarity with NoSQL databases (Cassandra, HBase, MongoDB).

- Exposure to data governance, compliance, and metadata management.

- Experience in containerized environments (Docker, Kubernetes).

- Understanding of machine learning pipelines and integration with data science workflows.


What Youll Get :

- Work on high-volume data infrastructure projects with global companies.

- Flexible remote work and performance-based culture.

- Opportunity to architect data solutions for cutting-edge applications.

- Access to the HYI.AI network for collaboration, growth, and career advancement.


info-icon

Did you find something suspicious?