Posted on: 03/03/2026
Note: Would prefer Female candidates for this role.
Description :
You will work closely with business stakeholders to understand requirements and deliver reliable cloud-based data products that support ongoing analytics and decision-making. Success in this role requires proven experience with cloud platforms, software development lifecycle, data engineering best practices, and data modeling.
Required Experience, Skills & Qualifications :
- Around 5 to 8 years of experience working with high-performance data products and large-scale data systems.
- Strong expertise in Python with hands-on experience in building and maintaining ETL pipelines, including data processing/manipulation libraries (e.g., pandas, PySpark, Dask).
- Proficiency in designing and developing scalable data pipelines and ETL processes, using tools such as AWS Glue, PySpark, Spark SQL, and orchestration frameworks like Airflow or AWS Step Functions.
- Expertise with AWS cloud services relevant to data engineering, including Glue, EMR, ECS, Lambda, Lake Formation along with other components for data processing and orchestration.
- Experience with databases across different paradigms (columnar, NoSQL, and MPP), such as Redshift, DynamoDB, Aurora, Postgres, and Snowflake.
- Proficiency in software engineering best practices, including version control (Git), containerization (Docker), unit/integration testing, and CI/CD pipelines.
- Strong data analysis skills with the ability to aggregate, transform, and prepare data for reporting and analytics.
- Knowledge of security, compliance, and design best practices for data solutions.
- Familiarity with API development and working with JSON/XML data formats.
- Proven ability to lead and mentor a team of Data Engineers to design, develop, and deliver data products.
- Excellent interpersonal, analytical, and communication skills for effective collaboration with stakeholders and cross-functional teams.
- Experience with reporting and visualization tools such as Tableau or Apache Superset is a plus.
Key Responsibilities :
- Build scalable end-to-end data pipelines to integrate and model datasets from diverse sources, ensuring alignment with functional and non-functional requirements.
- Translate business requirements and end-to-end designs into technical implementations that leverage system capabilities.
- Define and champion reusable, extensible, scalable, and maintainable solutions, while considering cost-benefit trade-offs.
- Conduct technical walk-throughs to ensure clear communication of system architecture.
- Collaborate with data engineering teams to deliver advanced cloud-based data products to clients.
- Interact with business and functional stakeholders to comprehend data requirements and downstream analytics needs.
- Validate technology solutions, produce concise design documentation, and contribute to work estimates.
- Cultivate a data-driven culture within the team and spearhead impactful data engineering projects.
- Stay informed about data engineering trends and integrate data best practices into software development, ensuring data integrity, scalability, and efficiency in alignment with the Roche motto : "Doing now what patients need next."
Good To Have Skills :
- Experience in the Healthcare Laboratory (IVD) domain is a plus.
- Experience with security and privacy regulations (GDPR, HIPAA, etc.)
- Demonstrated ability to collaborate effectively with cross-functional teams in a fast-paced and dynamic environment.
- Proven track record of conducting root cause analyses on both internal and external data and processes to address specific business inquiries and identify areas for enhancement.
The job is for:
Did you find something suspicious?
Posted by
Posted in
Data Engineering
Functional Area
Data Engineering
Job Code
1617692