Posted on: 30/10/2025
Description :
Were seeking a hands-on GenAI & Computer Vision Engineer with 35 years of experience delivering production-grade AI solutions.
You must be fluent in the core libraries, tools, and cloud services listed below, and able to own end-to-end model developmentfrom research and fine-tuning through deployment, monitoring, and iteration.
In this role, youll tackle domain-specific challenges like LLM hallucinations, vector search scalability, real-time inference constraints, and concept drift in vision models.
Key Responsibilities :
Generative AI & LLM Engineering :
- Fine-tune and evaluate LLMs (Hugging Face Transformers, Ollama, LLaMA) for specialized tasks.
- Deploy high-throughput inference pipelines using vLLM or Triton Inference Server.
- Design agent-based workflows with LangChain or LangGraph, integrating vector databases (Pinecone, Weaviate) for retrieval-augmented generation.
- Build scalable inference APIs with FastAPI or Flask, managing batching, concurrency, and rate-limiting.
Computer Vision Development :
- Develop and optimize CV models (YOLOv8, Mask R-CNN, ResNet, EfficientNet, ByteTrack) for detection, segmentation, classification, and tracking.
- Implement real-time pipelines using NVIDIA DeepStream or OpenCV (cv2); optimize with TensorRT or ONNX Runtime for edge and cloud deployments.
- Handle data challengesaugmentation, domain adaptation, semi-supervised learningand mitigate model drift in production.
MLOps & Deployment :
- Containerize models and services with Docker; orchestrate with Kubernetes (KServe) or AWS SageMaker Pipelines.
- Implement CI/CD for model/version management (MLflow, DVC), automated testing, and performance monitoring (Prometheus + Grafana).
- Manage scalability and cost by leveraging cloud autoscaling on AWS (EC2/EKS), GCP (Vertex AI), or Azure ML (AKS).
Cross-Functional Collaboration :
- Define SLAs for latency, accuracy, and throughput alongside product and DevOps teams.
- Evangelize best practices in prompt engineering, model governance, data privacy, and interpretability.
- Mentor junior engineers on reproducible research, code reviews, and end-to-end AI delivery.
Required Qualifications :
You must be proficient in at least one tool from each category below : .
LLM Frameworks & Tooling :
- Hugging Face Transformers, Ollama, vLLM, or LLaMA.
Agent & Retrieval Tools :
- LangChain or LangGraph; RAG with Pinecone, Weaviate, or Milvus.
Inference Serving :
- Triton Inference Server; FastAPI or Flask.
Computer Vision Frameworks & Libraries :
- PyTorch or TensorFlow; OpenCV (cv2) or NVIDIA DeepStream.
Model Optimization :
- TensorRT; ONNX Runtime; Torch-TensorRT.
MLOps & Versioning :
- Docker and Kubernetes (KServe, SageMaker); MLflow or DVC.
Monitoring & Observability :
- Prometheus; Grafana.
Cloud Platforms :
- AWS (SageMaker, EC2/EKS) or GCP (Vertex AI, AI Platform) or Azure ML (AKS, ML Studio).
Programming Languages :
- Python (required); C++ or Go (preferred).
Additionally :
- Bachelors or Masters in Computer Science, Electrical Engineering, AI/ML, or a related field.
- 35 years of professional experience shipping both generative and vision-based AI models in production.
- Strong problem-solving mindset; ability to debug issues like LLM drift, vector index staleness, and model degradation.
- Excellent verbal and written communication skills.
Typical Domain Challenges Youll Solve :
- LLM Hallucination & Safety : Implement grounding, filtering, and classifier layers to reduce false or unsafe outputs.
- Vector DB Scaling : Maintain low-latency, high-throughput similarity search as embeddings grow to millions.
- Inference Latency : Balance batch sizing and concurrency to meet real-time SLAs on cloud and edge hardware.
- Concept & Data Drift : Automate drift detection and retraining triggers in vision and language pipelines.
- Multi-Modal Coordination : Seamlessly orchestrate data flow between vision models and LLM agents in complex workflows.
Did you find something suspicious?