HamburgerMenu
hirist

Job Description

We Are Hiring : Python Developer

About the Role :

We're looking for a highly skilled and experienced Python Developer to join our team in Hyderabad or Bangalore.


You'll play a key role in designing, developing, and maintaining scalable data pipelines and high-performance applications. This is an exciting opportunity for someone with a strong background in big data technologies, cloud platforms, and modern API development to make a significant impact.

Work Location : Hyderabad / Bangalore (Hybrid - 3 days a week in office)

Experience : 8 - 12 Years

Key Responsibilities

- Develop and maintain robust applications : Design, build, and deploy efficient, reusable, and scalable code using Python.

- Build high-performance APIs : Create and manage RESTful APIs and microservices using FastAPI to support various business needs.

- Engineer data pipelines : Design and implement complex ETL pipelines using PySpark and Apache Spark for large-scale data processing.

- Manage big data processing : Write optimized Spark jobs and work with RDDs and DataFrames to handle data ingestion, cleansing, and transformation.

- Develop orchestration workflows : Utilize Apache Airflow to schedule, monitor, and orchestrate data workflows.

- Query and analyze data : Write and optimize complex SQL queries and integrate Spark SQL for data analysis and reporting.

- Collaborate with cloud technologies : Deploy and manage applications within a cloud environment (AWS, Azure, or GCP).

- Ensure code quality : Adhere to best practices for code development, testing, and documentation.

- Mentor and guide junior team members : Share knowledge and expertise with other developers on the team.

Mandatory Skills & Qualifications

- 8-12 years of professional experience as a Python Developer.

- Deep expertise in Python programming, including a strong understanding of its libraries, language features, and best practices.

- Extensive hands-on experience with Apache Spark and PySpark, including writing optimized Spark jobs, transformations, and actions.

- Proven experience with Apache Airflow for workflow orchestration.

- Solid experience developing high-performance RESTful APIs using FastAPI.

- Proficiency in SQL and experience with Spark SQL.

- Working knowledge of a major cloud platform such as AWS, Azure, or GCP.

- Strong understanding of big data processing concepts, including distributed computing and fault tolerance.

Desired Skills & Qualifications

- Familiarity with the Hadoop ecosystem (HDFS, Hive, etc.).

- Experience with various data serialization formats like JSON, Parquet, Avro, and ORC.

- Knowledge of software development lifecycle (SDLC) and Agile methodologies.

- Excellent problem-solving skills and the ability to work independently as well as part of a team.

- Strong communication and interpersonal skills.


info-icon

Did you find something suspicious?