Posted on: 18/12/2025
Description :
Position : Data Engineer with ETL & AI
Experience : 3-6+yrs
Location : Bangalore
Job Mode : Work From Office
Job Description :
- Collaborate closely with the reporting team, data scientists, product owners, and domain experts to design, develop, and maintain scalable data pipelines using an agile development methodology.
- Build efficient and reusable pipelines to ingest, process, and manage large-scale structured and unstructured datasets using cloud infrastructure, while ensuring compliance with data governance, lineage, and monitoring standards.
- Work effectively with diverse data sources, including structured datasets (e.g., RDBMS, Hive, Impala) and semi-structured datasets (e.g., JSON, XML, logs).
- Leverage cloud platforms and tools such as AWS, Dataiku, and Hadoop to architect and implement scalable, fault-tolerant data solutions.
- Automate and optimize ETL workflows to streamline data integration and transformation processes for enterprise-scale reporting and AI solutions.
- Develop and maintain a strong foundation in software engineering practices, including CI/CD pipelines, version control tools (e.g., Git), and advanced scripting with Python or PySpark (preferably).
- Utilize containerization and orchestration systems such as Jenkins, Dataiku, and Git to deploy, manage, and monitor scalable data solutions.
- Monitor, troubleshoot, and support operational data pipelines to ensure they meet user requirements, implementing bug fixes and necessary changes promptly.
- Collaborate with cross-functional teams to address requirements for data preparation and processing for advanced analytics, reporting, and machine learning models.
- Stay updated on emerging technologies, tools, and best practices in data engineering to improve data pipeline performance and scalability.
Your Profile :
You are an ideal candidate for this role if you have :
- A Bachelors or Masters degree in Computer Science, Mathematics, Statistics, or a related field.
- At least 3+ years of hands-on experience in data engineering, with a proven track record of designing and implementing robust data pipelines.
- Proficiency in working with both structured and unstructured data, leveraging databases (e.g., RDBMS, Hive, Impala), APIs and/or NoSQL systems (e.g. MongoDB).
- Strong experience with cloud services, preferably Dataiku, along with tools such as AWS and Hadoop, focused on building scalable and enterprise-grade pipelines.
- Solid expertise in automating and optimizing ETL workflows for both real-time and batch data processing.
- A strong foundation in software/data engineering practices, including CI/CD pipelines, version control tools (e.g., Git,Jenkins), and advanced scripting with Python or PySpark (preferably).
- Familiarity with deployment tools and platforms such as Jenkins to support production-ready pipelines.
- Hands-on experience with big data platforms (e.g., Spark, HDFS, Impala) and a strong understanding of data governance, lineage, and monitoring principles.
- Experience in supporting reporting, visualization, and data analytics tools such as Tableau, and collaborating with data scientists for feature engineering.
- Exceptional collaboration and communication skills, with experience working in cross-functional, agile teams.
- Excellent problem-solving abilities, analytical thinking, and attention to detail.
- A proactive, self-motivated, and results-oriented approach, with the ability to adapt and perform in a dynamic environment.
Good to have :
- Working knowledge of orchestration systems, including Docker, Kubernetes
- Knowledge of concepts around Data as a product, Data Mesh etc.
- Basics of ML and AI
Did you find something suspicious?
Posted by
Posted in
Data Engineering
Functional Area
Data Engineering
Job Code
1592302
Interview Questions for you
View All