HamburgerMenu
hirist

Auxo AI - Data Engineer - Google Cloud Platform

AuxoAI
Multiple Locations
7 - 11 Years

Posted on: 28/01/2026

Job Description

Description :


Auxo is looking for a Senior Data Engineer - GCP Native Platform to design, build, and operate scalable, production-grade data platforms for our enterprise clients.

This is a hands-on delivery role, not a support function. You will own end-to-end data pipeline engineering on Google Cloud Platform, embedding best-in-class data engineering practices across batch and streaming workloads, while working closely with architects, analysts, and data scientists.

Location : Bangalore, Hyderabad, Mumbai, and Gurgaon

Responsibilities :


- Design, build, and optimize scalable batch and streaming data pipelines using Dataflow (Apache Beam)

- Develop and manage workflow orchestration using Airflow on Cloud Composer

- Implement ELT transformations using Dataform for SQL-based data modeling and transformations

- Design and maintain BigQuery datasets following layered / medallion architecture patterns

- Implement event-driven ingestion and CDC patterns using Pub/Sub

- Partner with architects to implement technical designs, standards, and platform best practices

- Ensure performance optimization, reliability, monitoring, and cost efficiency of data pipelines

- Implement data quality checks, validations, and monitoring within pipelines

- Support production deployments, incident resolution, and operational stability

- Mentor junior engineers and contribute to engineering excellence across the team

Requirements :


Required Skills & Experience :

- Data Engineering (Strong Hands-on Experience)

- Design and development of production-grade data pipelines

- Batch and streaming data processing architectures

- Workflow orchestration and dependency management

- Data modeling, schema design, and performance optimization

- Pipeline monitoring, troubleshooting, and cost optimization

GCP Data Platform :

- Hands-on experience with BigQuery (advanced SQL, partitioning, clustering, optimization)

- Strong experience with Dataflow / Apache Beam (Python or Java)

- Experience with Cloud Composer / Airflow

- Experience with Pub/Sub and Cloud Storage

Technical Foundation :

- Strong proficiency in SQL and Python (Java is a plus)

- Solid understanding of ETL/ELT patterns and modern data stack concepts

- Experience with Git-based version control and CI/CD pipelines

- Working knowledge of cloud monitoring and logging

Preferred Qualifications :


- Experience with GCP Professional Data Engineer certification

- Exposure to Dataform or dbt for transformation workflows

- Experience with real-time streaming architectures

- Familiarity with Vertex AI, Cloud Functions, or Dataproc

- Understanding of data governance concepts and platforms (Dataplex, Atlan, Collibra)

- Experience with legacy-to-cloud data migrations

- Familiarity with Looker or Power BI


info-icon

Did you find something suspicious?

Similar jobs that you might be interested in