ritishsajjagcp (1)
ritishsajjagcp (1)
Passionate data scientist with expertise in training, evaluating and deploying deep learning
models with Tensorflow/ Keras/ PyTorch. Experience in classical machine learning(XGBoost,
random forest) and handling complex data for deep learning (instance/semantic
segmentation/object detection).
Version control
Git, GitHub, Git Lab, Bitbucket
tool
Programming
Python, C++, HTML5, MYSQL,
Languages
SKILLS:
PROFESSIONAL EXPERIENCE:
Senior Data Scientist/Machine Learning:
Morgan Stanley, NY June 2022 -Oct 2024
• Deployed Azure IaaS virtual machines and cloud services into secure VNets and
subnets. Managed and optimized CI/CD pipelines using Azure DevOps for seamless
build and release automation.
• Prepared, trained, deployed, and monitored machine learning models using Azure
Pipelines, ensuring robust model performance and addressing data drift issues.
• Designed reproducible training workflows to reduce variability across model iterations.
• Implemented LLM-based chatbots (using OpenAI GPT, LLAMA2, Mistral
models) integrated with Janus Graph and Gremlin queries, enhancing user
interactions and customer service.
• Developed and deployed AI chatbots using Large Language Models like OpenAI’s
GPT and Hugging Face models.
• Leveraged Databricks Asset Bundles (DAB), Delta Lake, and MLflow for scalable
machine learning pipeline management, reducing downtime and increasing efficiency
by 16%.
• Authored Terraform scripts to automate the deployment of Azure cloud services,
creating reusable templates for multi-tier applications and provisioning cloud
infrastructure.
• Set up Jenkins pipelines integrated with tools like GIT, Nexus, SonarQube, Ansible,
and Docker. Configured additional Docker Slave Nodes for Jenkins CI/CD using
custom Docker Images.
• Worked extensively with Docker for containerized deployments, including Docker
Hub, Docker Compose, Docker Weave, and Docker Trusted Registry.
• Implemented container orchestration with Kubernetes, integrating EFK stack
(Elasticsearch, Fluentd, Kibana) for logging and Prometheus with Grafana for
cluster monitoring and alerting.
• Enforced network policies in K3s clusters using Calico CNI and explored CNCF
container runtimes for performance benchmarking.
• Configured and managed tools like Splunk, Nagios, CloudWatch, and the ELK stack
for system monitoring, log analysis, and visualizations.
• Integrated Adobe Analytics, Conviva, and Datadog to monitor platform performance,
generating insights to enhance stability.
• Conducted benchmarking using tools like Sysbench, JMeter, and Apache Bench to
evaluate container and orchestration platform performance.
• Built connectors for databases, APIs, and web scraping tools to integrate real-time
and historical data into LLM workflows.
• Regularly collaborated with business teams to finalize requirements, define model
monitoring metrics, and review implementation plans in an Agile environment.
Data Analyst:
June 2015 - February 2017 SetuServ informatics, India
● Worked with requirements analysts and subject matter experts to identify, understand,
and document business needs for the data flow.
● Worked in close coordination with a variety of Financial, Mortgage, and Credit
Consumer Group business teams in gathering the business requirements.
● Worked with Central Distribution Hub (CDH) team in developing strategies to handle
data from EO (Enterprise Originations) to CDH, then CDH to downstream systems.
● I used the export-import feature to carry out day-to-day migrations of various ETL
Informatica objects.
● I worked with Chief Data Architects to slice the data requirements into work streams and
various components.
● Data mapping, logical data modeling, class diagrams, and ER diagrams were prepared
with the SQL queries and the PL/SQL stored procedures to filter data within the Oracle
database.
● We performed development tasks in ETL Informatica, such as job creation using
different stages, debugging, etc.
● It is developed and maintained a data dictionary to create metadata reports for both
technical and business purposes.
● A resource for analytical services using SQL Server, TOAD/Oracle.
● Execute the following SQL and PL/SQL queries using TOAD and SQL Navigator:.
● Identify and document the transformation rules and identified/documented data sources
needed to populate and maintain data warehouse content.
● I created the many KPIs required for the Dashboards in Excel and Cognos using
formulas, variables, standard business object functions, and merging between multiple
universes fetching information from different underlying data targets.
● Developed and implemented basic SQL queries for testing and data validation reports.
Made use of data warehousing for profiling the data available in an existing database.