HamburgerMenu
hirist

Data Engineer - Pentaho/ETL Tools

Reflion Tech Private Limited
Multiple Locations
3 - 10 Years

Posted on: 16/12/2025

Job Description

Experience : 3 to 6 years in Pentaho, overall could be more

Location : Permanent Remote, India

Salary : Based on fitment

Notice Period : 30 days preferred

Overview :



Our client is modernizing it's data integration landscape and are seeking a skilled Pentaho Developer / Data Engineer who is excited about transforming legacy ETL systems into scalable, cloud ready data pipelines using industry-leading technologies.

You will help maintain and enhance existing Pentaho Data Integration (PDI/Kettle) processes while playing a key role in our migration to modern data engineering platforms. You will collaborate with technical and business teams across our portfolio to ensure data flows are robust, efficient, and future ready.

This role is ideal for a developer / data engineer who enjoys solving complex data problems and is passionate about modernization, performance, and continuous improvement.

Key Responsibilities :


- Develop, maintain, and optimize ETL workflows using Pentaho Data Integration (PDI/Kettle).

- Document and analyze existing Pentaho jobs, data flows, dependencies, and performance bottlenecks.

Contribute to and execute the migration strategy from Pentaho to modern ETL/data integration platforms, such as :

- Talend / Talend Cloud

- Informatica Cloud (IICS)

- Azure Data Factory

- AWS Glue

- Dbt (Data Build Tool)

- Snowflake pipelines (Tasks, Streams, Snowpipe)

- Work with architects and engineering leads to shape target-state data integration architecture.

- Implement data validation, quality checks, and reconciliation processes.

- Assist in building scalable, maintainable, and secure data pipelines across on-prem and cloud environments.

- Support production workloads, resolve incidents, and ensure reliability of critical data processes.

- Produce clear technical documentation and migration runbooks.

Required Qualifications :



- 3 to 6+ years of hands-on experience with Pentaho PDI/Kettle or similar ETL platforms.

- Strong SQL development skills and solid understanding of relational database concepts.

- Experience with at least one modern ETL/cloud data integration tool, including :

- Talend

- Informatica Cloud (IICS)

- Azure Data Factory

- AWS Glue

- Understanding of cloud ecosystems (Azure, AWS, or GCP).

- Familiarity with Git-based version control, CI/CD, and ETL/job scheduling tools.

- Excellent problem-solving skills, attention to detail, and ability to work collaboratively in a distributed team.

Preferred Skills :



- Experience with modern cloud data warehouses: Snowflake, BigQuery, Redshift, etc.

- Exposure to data modeling (dimensional, star schema, SCD patterns).

- Python scripting for automation and data engineering tasks.

- Experience supporting ETL modernization or cloud migration projects.


info-icon

Did you find something suspicious?