Posted on: 06/11/2025
Job Title : Senior Data Engineer Data Platform
Hybrid : 3 days WFO in a week
Office Location : Mumbai - Commerz II - Goregaon(East)
Work Timings : 03 : 00 pm to midnight
Our customers and AI agents are pushing the limits of modern data warehouse evolution. Be part of it. Youll contribute to architecting the next generation of our data platformbalancing real-time serving (<100ms cache hits), analytical queries (<5s aggregates), and streaming ingestion (<2min to validated), all while serving institutional investors analyzing markets 10+ hours daily.
EDS Track Record : 8 years profitable. Northern Trust partnership (investor & channel). SOC 2 Type II certified. Tier-1 hedge fund clients. Transparent runway reviewed in town halls. Leadership that still writes code.
The Role in Reality :
Current State : Snowflake data warehouse with batch processing, dbt transformations, Redis caching layer.
Core Ownership : This role owns denormalized materialized datasets end-to-endfrom contract-safe inputs to deterministic, low-latency outputs.
What Youll Build (Future Direction) :
- Streaming pipelines (Kafka/Flink/Snowflake) replacing batch processes
- Enhanced Snowflake lakehouse with contract-enforced stage promotion (raw/validated/modeled)
- Redis-fronted serving layer backed by ClickHouse/StarRocks for real-time queries
- Expanded governance with JSON Schema contracts, Great Expectations validation, Airflow orchestration
Latency & SLOs (Truthful) :
- Hot path (Redis over ClickHouse) : p95 <100ms cache hits; <700ms cache-miss
- Historical analytics (Snowflake) : p95 <5s for defined aggregates; ad-hoc queries out of SLO
- Ingestion (Kafka/Flink/validated) : p95 <2m end-to-validated
- Error budget : 0.5% monthly (~3.6 hours downtime allowed)
How We Ship : Each pipeline ships with SLIs, SLOs, runbook, and an error budget. Canary every DAG change; rollback via table version. All infra through Terraform / AWS CDK; PRs require docs, owners, tests; backfills are first-class with quotas.
Youll help us evolve from reliable batch to real-time streaming while maintaining production stability.
We evaluate work samples, not algorithmic puzzles. Our engineering culture values sustainable urgencyship quickly but thoughtfully.
Required Experience (Be Ready to Discuss Specific Examples) :
Data Platform Engineering :
- Snowflake at scale : Share a query optimization with before/after metrics and cost impact
- Batch & streaming concepts : Describe any streaming/real-time data challenge youve solved (any technology)
- Performance optimization : Explain how you improved data pipeline latency or throughput in production
Reliability & Operations :
- Data quality : Walk through data validation or quality controls youve implemented
- Incident response : Detail a data incident from detection to resolution
- Observability : Show monitoring/alerting you created that prevented customer impact
Technical Fundamentals :
- SQL optimization : Complex query you rewrote for significant performance improvement
- Python at scale : Data processing code youve optimized for memory/speed
- Data engineering patterns : Implementation of idempotency, deduplication, or consistency in production
Interview focus : Well explore your production experiences and how they apply to our evolution.
Preferred Experience :
Did you find something suspicious?
Posted By
Shehzarin
Hr Head at EQUITY DATA SCIENCE INDIA PRIVATE LIMITED
Last Active: NA as recruiter has posted this job through third party tool.
Posted in
Data Engineering
Functional Area
Data Engineering
Job Code
1570875
Interview Questions for you
View All