- Bachelor's or Masters degree in Computer Science, Data Science, Statistics, or a related field
- 6+ years of experience in Data Scientist and relevant stream
- Hands-on experience with AutoML platforms and deep learning frameworks
- Familiarity with synthetic data generation tools and their applications
- Solid programming skills in Python, R, and SQL
- Solid understanding of statistics, mathematics, and data modeling techniques
- Proven excellent problem-solving skills and the ability to work independently or in a team
- Proven solid communication and documentation skills.
Primary Responsibilities :
- Analyze large and complex datasets to extract actionable insights and support data-driven decision-making
- Design, develop, and deploy machine learning and deep learning models for various business use cases
- Utilize AutoML tools (e.g., H2O.ai, Google Cloud AutoML, DataRobot) to accelerate model development
- Apply classical ML algorithms such as Logistic Regression, Decision Trees, Clustering (K-means, Hierarchical, SOM), PCA, Bayesian models, and Time Series forecasting (ARIMA/ARMA)
- Implement deep learning techniques including CNNs, LSTMs, GRUs, and optimization methods like Adam, Adagrad, and regularization techniques (L1, L2)
- Work on recommender systems using Collaborative Filtering, FPMC, FISM, Fossil, etc.
- Develop and optimize models for NLP, speech, and image processing applications
- Use deep learning frameworks such as TensorFlow, Keras, PyTorch, XGBoost, Caffe, and Theano for model development and deployment
- Generate synthetic data using tools like Gretel.ai and Synthea when real data is limited or sensitive
- Collaborate using tools like Confluence for documentation and knowledge sharing
- Stay updated with the latest research and advancements in AI/ML and apply them to real-world problems