Posted on: 01/12/2025
Key Responsibilities :
Python Proficiency :
- Demonstrate a strong command of Python programming language, actively contributing to the development and maintenance of data engineering solutions.
Data Engineering Expertise :
- Set up and maintain efficient data pipelines, ensuring smooth data flow and integration across. systems.
- Experience with PySpark and/or Data Bricks platform is required
- Contribute to the establishment and maintenance of Data Lakes, implementing industry best practices.
Execute data scrubbing techniques and implement data validation processes to ensure data integrity and quality.
Machine Learning Integration :
- Experience in setting up and managing classic ML and LLM models.
- Participate in model updates and validation, contributing to the enhancement of machine learning algorithms.
Tool and Platform Proficiency :
- Experience with PySpark and/or Data Bricks platform is required, apart from this expertise in at least one popular tool/platform within the data engineering domain will be nice to have.
- Stay informed about industry trends, exploring and adopting tools to optimize data engineering processes.
Collaboration and Communication :
- Collaborate effectively with cross-functional teams, including data scientists, software engineers, and business analysts.
- Communicate technical concepts to non-technical stakeholders, fostering collaboration and innovation within the team.
Documentation and Best Practices :
- Contribute to maintaining comprehensive documentation for data engineering processes, ensuring knowledge transfer within the team.
- Adhere to best practices in data engineering, promoting a culture of quality and efficiency.
Must-Have Skills :
- Bachelors or Masters degree in Computer Science, Data Science, or a related field.
- Minimum of 2 years of proven experience as a Data Engineer.
- Experience with PySpark.
- Strong proficiency in Python programming language.
- Experience in setting up and managing data pipelines, data lakes, and classic Machine Learning models.
- Familiarity with at least one popular tool/platform in the data engineering space (Kafka,Airflow, Snowflake, etc.
Nice-to-Have Skills (Optional) :
- Exposure to cloud-based data platforms (AWS/Azure/GCP).
- Basic knowledge of big data technologies such as Hadoop or Spark.
- Familiarity with containerization tools like Docker or Kubernetes.
- Interest in data visualization tools (Tableau, PowerBI, etc.
- Certifications in relevant data engineering or machine learning technologies.
Did you find something suspicious?
Posted By
Posted in
Data Engineering
Functional Area
Data Engineering
Job Code
1583198
Interview Questions for you
View All