HamburgerMenu
hirist

Principal Data Engineer - Big Data Technologies

K&K
Anywhere in India/Multiple Locations
10 - 14 Years

Posted on: 13/10/2025

Job Description

Description :


Title : Principal Data Engineer


Location : Bangalore, Bhopal, Chennai, Gurugram, Hyderabad, Jaipur, Pune


Mode of Operation : Hybrid


Employment Type : Full-time Permanent


Experience : 10 -14 years


About the Role :


We are seeking an accomplished and highly technical Principal Data Engineer to join our emerging Data Warehouse team across multiple locations.


This role is a critical leadership position that requires deep expertise in cloud-native data architecture, advanced data modeling, and practical delivery using modern tools like Spark, Databricks, and dbt.


As a Principal Engineer, you will be relied upon to lead a team of senior engineers, define the technical roadmap, manage client interactions, and solve the most complex data challenges in building our innovative BNL Data Warehouse offerings.


Responsibilities :


- Technical Leadership & Architecture : Act as the technical authority in the Data Warehouse team, leading a team of 4-5 Senior Data Engineers and taking ownership of proposing, reviewing, and defining scalable, secure, and cost-efficient data architectures across cloud environments (GCP or AWS).


- Advanced Data Engineering : Design, develop, and maintain high-performance, resilient data pipelines using Python, PySpark, and JVM languages, ensuring reliability and efficiency at petabyte scale.


- Modern Data Warehouse Development : Lead the design and development of data warehouses, implementing advanced data models (e.g., Data Vault, Kimball) suitable for the business nature, utilizing modern cloud data warehouses, and ensuring data quality and governance.


- Orchestration & Workflow Management : Possess high proficiency in architecting and implementing robust workflow management systems, with Airflow expertise being mandatory for defining complex ETL/ELT orchestration.


- Databricks and Transformation : Drive data transformation strategy using Databricks and proficiency with dbt (data build tool) for managing and deploying analytical data transformations with software engineering rigor.


- Complex Problem Solving : Solve the most complex situations and data models, ensuring optimal data organization and query performance for downstream analytics and marketing suites.


- Client Engagement & Strategy : Interact directly with clients for requirement gathering, translating ambiguous business needs into clear, scalable technical solutions, and building a comprehensive, holistic roadmap for feature delivery.


- Quality & CI/CD : Lead the integration of advanced CI/CD processes into the data pipeline lifecycle, with strong experience in building automated pipelines, experience with GitHub Actions, ArgoCD, or Jenkins is highly valued.


- Risk Management & Delivery : Proactively identify technical and project risks and develop effective mitigation strategies to ensure the timely and high-quality delivery of features, always maintaining a focus on business value over engineering interest.


Required Experience :


- Overall Experience : 10+ years of total experience in the field of data engineering, with 4+ years experience in a technical leadership or team lead role.


- Cloud Native Expertise : 5+ years of proven experience building and operating cloud-native, data-intensive applications on either GCP or AWS.


- Programming & Scripting : 8+ years of hands-on, production experience using Python, JVM languages (Java/Scala), and Shell Scripting.


- Big Data Processing : Mandatory, hands-on experience building and maintaining high-volume Spark applications (using PySpark or Scala/Java) for large-scale data processing.


- Architecture & IaC : Mandatory proficiency on an IaC tool such as Terraform, AWS CDK, or CloudFormation, and a proven ability to propose and review detailed cloud architecture.


- Data Lake Fundamentals : Fundamental understanding of efficient file formats like Parquet, Delta Lake, and other Open Table Formats (OTFs).


- Testing and Automation : Strong experience with unit testing and test automation frameworks specific to data pipelines and applications.


- Observability : Experience implementing and utilizing observability systems (such as Datadog or Prometheus/Grafana) for pipeline health, monitoring, and alerting.


Preferred Skills :


- Experience with containerization and orchestration technologies (e.g., Docker, Kubernetes).


- Prior experience with specialized marketing suite data integrations.


- Experience implementing Data Mesh or other advanced data governance paradigms.


- Excellent documentation skills with the ability to provide good quality, readable business documentation in addition to technical specifications.


- Proactive and service-oriented approach when interacting with internal customers and stakeholders.


info-icon

Did you find something suspicious?