HamburgerMenu
hirist

Job Description

Data Engineer - Python

Experience : 5+ Years

Employment Type : Full-time

Job Overview :


We are seeking a highly skilled Data Engineer with 5+ years of experience, specializing in Python, to join our team. You'll be instrumental in designing, building, and maintaining robust and scalable data pipelines, ensuring data quality, and enabling data-driven decision-making. This role requires strong proficiency in Python for data manipulation, ETL processes, and integration within various data ecosystems.

Key Responsibilities :


- Design, develop, and optimize scalable data pipelines using Python to extract, transform, and load (ETL/ELT) data from diverse sources into data warehouses or data lakes.

- Write clean, efficient, and well-documented Python code for data processing, automation, and integration tasks.

- Collaborate with data scientists, analysts, and other engineering teams to understand data requirements and translate them into robust technical solutions.

- Implement best practices for data quality, data governance, and data integrity throughout the data lifecycle.

- Monitor, troubleshoot, and optimize existing data pipelines for performance, reliability, and cost-efficiency.

- Develop and maintain automated processes for data validation, reconciliation, and error handling.

- Work with various database technologies and big data platforms (e.g., SQL, NoSQL, data warehouses, cloud data services).

- Contribute to the continuous improvement of data engineering processes, tools, and methodologies.

Required Skills :


- 5+ years of professional experience as a Data Engineer.

- Expert-level proficiency in Python for data manipulation, scripting, and building data pipelines.

- Strong experience with ETL/ELT concepts and tools/frameworks in a Python ecosystem (e.g., Pandas, Dask, PySpark, Apache Airflow).

- Solid understanding and hands-on experience with relational databases (SQL) and data warehousing

principles.

- Experience with cloud platforms and their data services (e.g., AWS, Azure, GCP - including services like S3, Redshift, Glue, Data Factory, BigQuery, etc.).

- Familiarity with version control systems like Git.

- Strong analytical and problem-solving skills, with an eye for detail.

- Excellent communication and collaboration skills to work effectively within cross-functional teams.

Preferred Qualifications :


- Experience with Big Data technologies (Apache Spark, Hadoop, Kafka).

- Hands-on experience with NoSQL databases (MongoDB, Cassandra, DynamoDB).

- Familiarity with data modeling techniques (dimensional modeling, 3NF).

- Experience with CI/CD pipelines for data solutions.


info-icon

Did you find something suspicious?