HamburgerMenu
hirist

Snowflake Data Engineer

EduRun Group
Bangalore
10 - 12 Years

Posted on: 03/10/2025

Job Description

Job Description :


Key Responsibilities :


- Design, build, and optimize scalable batch and real-time streaming data pipelines using Databricks (PySpark) and Snowflake.

- Lead the architecture and implementation of application data stores using PostgreSQL, DynamoDB, and advanced SQL to support diverse application needs.

- Develop efficient data processing workflows using SQL, Python, and PySpark across structured (CSV, SQL), semi-structured (JSON, XML), and unstructured (PDF, logs) data formats.

- Utilize deep knowledge of Azure and AWS data ecosystems, including Microsoft Fabric and distributed computing frameworks, to design robust data solutions.

- Architect low-latency, high-throughput data pipelines using Apache Spark and cloud-native services ensuring high availability and fault tolerance.

- Implement CI/CD pipelines, automation scripts, schema versioning, and enforce rigorous data security and governance policies in production environments.

- Ensure best practices around data performance, scalability, compliance, and cost optimization are strictly followed.

- Collaborate with data scientists, analysts, and cross-functional teams to deliver reliable and scalable data solutions.


- Continuously evaluate and recommend improvements in data architecture and processes to support evolving business needs.

Required Skills & Qualifications :


- 10- 12 years of professional experience as a Data Engineer with proven expertise in Snowflake and Databricks (PySpark).

- Strong programming skills in SQL, Python, and PySpark for complex data transformations and pipeline development.

- Extensive experience working with various data formats : structured, semi-structured, and unstructured.

- In-depth knowledge of PostgreSQL and DynamoDB for designing and managing application data stores.

- Strong understanding of cloud data ecosystems in Azure and AWS, including Microsoft Fabric and distributed computing frameworks.

- Experience designing and maintaining scalable batch and streaming pipelines with a focus on low latency and high throughput.

- Familiarity with CI/CD automation, schema versioning, and production-grade data security practices.

- Ability to apply best practices in data engineering including performance tuning, data governance, and compliance standards.

- Excellent problem-solving skills and ability to work collaboratively in a team environment.

Preferred Qualifications :


- Experience with additional cloud services and data platforms.

- Prior exposure to data mesh architectures or data lakehouse concepts.

- Certifications in Snowflake, Databricks, Azure, or AWS data engineering.


info-icon

Did you find something suspicious?