Posted on: 04/12/2025
Description :
The Opportunity : Building the Data Foundation for AI
Detail Specification :
Experience : 3 - 6 Years
Location : Bangalore, Mumbai, or Gurgaon (Hybrid / Work from Office)
Mandatory Background : B.Tech/M.Tech from IIT, NIT, BITS Pilani, or IIIT
The Company : High-Growth B2B AI SaaS
About the Role :
We are seeking a Senior Data Engineer to design, build, and optimize our core data infrastructure. As the foundation of an AI-driven product, data quality and availability are non-negotiable. You will be responsible for creating robust, scalable ETL/ELT pipelines that ingest massive, complex datasets from enterprise clients, transform them for analytics and machine learning, and store them efficiently in our cloud data warehouse. This role is critical for the success of our predictive and generative AI models.
Key Technical Responsibilities :
1. Data Pipeline Architecture :
- Design & Implementation : Architect and implement highly resilient and scalable ETL/ELT pipelines using modern orchestration tools like Apache Airflow, Dagster, or Prefect.
- Big Data Processing : Utilize distributed processing frameworks (Spark or similar) for high-volume data transformations and cleansing processes.
- Data Lake/Warehouse : Design the schema, implement data modeling strategies, and optimize performance for our cloud data warehouse (Snowflake, Google BigQuery, or AWS Redshift).
2. Data Quality and Governance :
- Ingestion : Build connectors and APIs to securely ingest data from diverse enterprise sources (databases, APIs, streaming sources) in a compliant manner.
- Data Quality : Implement data validation and monitoring frameworks to proactively detect and alert on data quality issues, ensuring accuracy and consistency for ML models and business reporting.
- Performance : Optimize data pipelines for cost efficiency and low latency, ensuring timely data delivery for real-time analytics and model inference.
3. ML and Analytics Enablement :
- ML Data Prep : Work directly with Data Scientists/ML Engineers to define and provide the necessary structured and feature-engineered datasets for model training and evaluation.
- BI & Reporting : Ensure the data warehouse is structured efficiently to support rapid querying and dashboard development by the Business Intelligence team.
What You'll Bring (Mandatory Skills & Experience) :
- Educational Excellence : B.Tech/M.Tech in Computer Science or a related discipline from an IIT, NIT, BITS Pilani, or IIIT is mandatory.
- Experience : 3-6 years of professional experience in Data Engineering, specializing in building and scaling production-grade data pipelines.
- Core Skills : Expert proficiency in SQL and Python (specifically for data manipulation and scripting).
- Tools : Strong practical experience with a cloud data warehouse (Snowflake/BigQuery/Redshift) and a workflow orchestrator (Airflow preferred).
- Cloud : Familiarity with data services and infrastructure on a major cloud platform (AWS or GCP).
Logistics and Compensation :
- Location : This is a Hybrid Work / Work from Office Only position in Bangalore. No remote options are available.
Did you find something suspicious?
Posted By
Posted in
Data Engineering
Functional Area
Data Engineering
Job Code
1584814
Interview Questions for you
View All