Posted on: 16/07/2025
Job Description :
We are seeking immediate joiners (available to start in May) for the role of Big Data Developer (Python / PySpark).
The ideal candidate will have 58 years of industry experience, with at least 3+ years of hands-on work in Big Data technologies.
You will play a key role in building and maintaining scalable data pipelines, supporting streaming data use cases, and collaborating with business and engineering teams on architectural and technical decisions.
Key Responsibilities :
- Design, develop, and maintain scalable Big Data solutions using Python and PySpark.
- Build and optimize data pipelines to process large volumes of structured and unstructured data.
- Work with streaming services such as Kinesis, Kafka, or equivalent to support near real-time data processing.
- Integrate with NoSQL data stores such as HBase, DynamoDB, etc.
- Develop data processing jobs using Hadoop ecosystem tools such as Spark, Hive, Nifi, Flink, and Spark Streaming.
- Write complex SQL queries and leverage analytical SQL functions for data analysis and transformations.
- Participate in architectural discussions and make key technical decisions to improve data platform performance and reliability.
- Collaborate with cross-functional teams to understand business requirements and deliver high-quality data solutions.
Required Skills & Experience :
- 5 - 8 years of overall industry experience.
- Minimum 3+ years of hands-on experience in Big Data technologies.
- Strong programming experience in Python and PySpark; familiarity with Scala is a plus.
- Experience with Big Data streaming services such as Kinesis, Kafka, or similar.
- Experience with NoSQL data stores such as HBase, DynamoDB, etc.
- Solid understanding and hands-on experience with the Hadoop ecosystem, including Spark, Hive, Nifi, Flink, Spark Streaming, etc.
- Strong SQL skills, including writing complex queries and using analytical functions.
- Proven experience participating in business, architectural, and technical decision-making.
Nice to Have :
- Experience with AWS, Azure, or other cloud platforms.
- Familiarity with CI/CD pipelines for data engineering.
- Exposure to data security, governance, and compliance best practices.
Behavioural Competencies :
- Strong problem-solving and analytical skills.
- Excellent communication and collaboration abilities.
- Ability to work independently and in a team-oriented, collaborative environment.
- Ownership mindset with the ability to deliver under tight timelines
Did you find something suspicious?
Posted By
Posted in
Data Engineering
Functional Area
Big Data / Data Warehousing / ETL
Job Code
1513964
Interview Questions for you
View All