HamburgerMenu
hirist

Job Description

Job description :

We are a team of data professionals who work with large datasets every day. With over 40 years of combined experience, we have a unique approach to data engineering.

Key Responsibilities :

- Design and develop big data pipelines using Azure Databricks, PySpark, and Delta Lake.

- Work with Azure Data Lake, Azure Synapse, Azure SQL Database, and Azure Data Factory to implement robust data solutions.

- Develop and maintain ETL/ELT workflows for efficient data ingestion, transformation, and processing.

- Implement data governance, security, and compliance best practices in Azure environments.

- Optimize Databricks clusters, workflows, and cost efficiency in cloud environments.

- Collaborate with data scientists, analysts, and business stakeholders to ensure high-quality data solutions.

- Implement CI/CD pipelines for data engineering workflows using Azure DevOps.

Required Qualifications & Skills :

- Databricks Certified Data Engineer Associate (Preferred)

- Databricks Certified Data Engineer Professional

- Azure Cloud Services : Azure Databricks, Azure Data Factory, Azure Data Lake, Azure Synapse, Azure Functions

- Big Data & ETL : PySpark, SQL, Delta Lake, Kafka (Preferred)

- Programming : Python, SQL, Scala (Optional)

- Orchestration & Automation : Airflow, Azure DevOps, GitHub Actions

- Data Governance & Security : Unity Catalog, RBAC, PII masking

- Performance Optimization : Spark tuning, Databricks cluster configuration

- Experience : years of experience in data engineering with a focus on Azure and Databricks.

- Strong understanding of data modeling, warehousing, and governance best practices


info-icon

Did you find something suspicious?