Job Description

Work From Office only : Jaipur, Rajasthan.

Must have experience : 4 year+.

Should be strongly skilled in FastAPI, RAG, LLM, Generative AI.

About the Role :

We are seeking a hands-on and experienced Data Scientist with deep expertise in Generative AI to join our AI/ML team.

You will be instrumental in building and deploying machine learning solutions, especially GenAI-powered applications.

Key Responsibilities :

- Design, develop, and deploy scalable ML and GenAI solutions using LLMs, RAG pipelines, and advanced NLP techniques.

- Implement GenAI use cases involving embeddings, summarization, semantic search, and prompt engineering.

- Fine-tune and serve LLMs using frameworks like vLLM, LoRA, and QLoRA; deploy on cloud and on-premise environments.

- Build inference APIs using FastAPI and orchestrate them into robust services.

- Utilize tools and frameworks such as LangChain, LlamaIndex, ONNX, Hugging Face, and VectorDBs (Qdrant, FAISS).

- Collaborate closely with engineering and business teams to translate use cases into deployed solutions.

- Guide junior team members, provide architectural insights, and ensure best practices in MLOps and model lifecycle.

- Stay updated on latest research and developments in GenAI, LLMs, and NLP.

Required Skills and Experience :

- 4-8 years of hands-on experience in Data Science/Machine Learning, with a strong focus on NLP and Generative AI.

- Proven experience with LLMs (LLaMA 1/2/3, Mistral, FLAN T5) and concepts like RAG, fine-tuning, embeddings, chunking, reranking, and prompt optimization.

- Experience with LLM APIs (OpenAI, Hugging Face) and open-source model deployment.

- Proficiency in LangChain, LlamaIndex, and FastAPI.

- Understanding of cloud platforms (AWS/GCP) and certification in a cloud technology is preferred.

- Familiarity with MLOps tools and practices for CI/CD, monitoring, and retraining of ML models.

- Ability to read and interpret ML research papers and LLM architecture diagrams.

info-icon

Did you find something suspicious?