Anshul Jindal
LLMOps-driven fine-tuning, evaluation, and inference with NVIDIA NIM & NeMo Microservices
#1about 6 minutes
Understanding the GenAI lifecycle and its operational challenges
The continuous cycle of data processing, model customization, and deployment for GenAI applications creates production complexities like a lack of standardized CI/CD and versioning.
#2about 2 minutes
Breaking down the structured stages of an LLMOps pipeline
An effective LLMOps process moves a model from an experimental proof-of-concept through evaluation, pre-production testing, and finally to a production environment.
#3about 4 minutes
Introducing the NVIDIA NeMo microservices and ecosystem tools
NVIDIA provides a suite of tools including NeMo Curator, Customizer, Evaluator, and NIM, which integrate with ecosystem components like Argo Workflows and Argo CD for a complete LLMOps solution.
#4about 4 minutes
Using NeMo Customizer and Evaluator for model adaptation
NeMo Customizer and Evaluator simplify model adaptation through API requests that trigger fine-tuning on custom datasets and benchmark the resulting model's performance.
#5about 3 minutes
Deploying and scaling models with NVIDIA NIM on Kubernetes
NVIDIA NIM packages models into optimized inference containers that can be deployed and auto-scaled on Kubernetes using the NIM operator, with support for multiple fine-tuned adapters.
#6about 4 minutes
Automating complex LLM workflows with Argo Workflows
Argo Workflows enables the creation of automated, multi-step pipelines by stitching together containerized tasks for data processing, model customization, evaluation, and deployment.
#7about 3 minutes
Implementing a GitOps approach for end-to-end LLMOps
Using Git as the single source of truth, Argo CD automates the deployment and management of all LLMOps components, including microservices and workflows, onto Kubernetes clusters.
#8about 3 minutes
Demonstrating the automated LLMOps pipeline in action
A practical demonstration shows how Argo CD manages deployed services and how a data scientist can launch a complete fine-tuning workflow through the Argo Workflows UI, with results tracked in MLflow.
Related jobs
Jobs that call for the skills explored in this talk.
Full Stack Developer (all genders welcome)
ROSEN Technology and Research Center GmbH
Osnabrück, Germany
Senior
Matching moments
06:19 MIN
Defining LLMOps and understanding its core benefits
From Traction to Production: Maturing your LLMOps step by step
39:32 MIN
Implementing a CI/CD pipeline for your NLP model
Multilingual NLP pipeline up and running from scratch
01:01 MIN
Understanding the role and challenges of MLOps
The Road to MLOps: How Verivox Transitioned to AWS
40:05 MIN
How to assess and advance your LLMOps maturity
From Traction to Production: Maturing your LLMOps step by step
09:27 MIN
Using MLOps infrastructure to implement model governance
Model Governance and Explainable AI as tools for legal compliance and risk management
01:58 MIN
The convergence of ML and DevOps in MLOps
AI Model Management Life Circles: ML Ops For Generative AI Models From Research to Deployment
10:29 MIN
What MLOps is and the engineering challenges it solves
MLOps - What’s the deal behind it?
05:43 MIN
Introducing NVIDIA NIM for simplified LLM deployment
Efficient deployment and inference of GPU-accelerated LLMs
Featured Partners
Related Videos
DevOps for AI: running LLMs in production with Kubernetes and KubeFlow
Aarno Aukia
Adding knowledge to open-source LLMs
Sergio Perez & Harshita Seth
Efficient deployment and inference of GPU-accelerated LLMs
Adolf Hohl
From Traction to Production: Maturing your LLMOps step by step
Maxim Salnikov
MLOps on Kubernetes: Exploring Argo Workflows
Hauke Brammer
MLOps - What’s the deal behind it?
Nico Axtmann
Effective Machine Learning - Managing Complexity with MLOps
Simon Stiebellehner
Self-Hosted LLMs: From Zero to Inference
Roberto Carratalá & Cedric Clyburn
Related Articles
View all articles.gif?w=240&auto=compress,format)
.gif?w=240&auto=compress,format)
.gif?w=240&auto=compress,format)
.png?w=240&auto=compress,format)
From learning to earning
Jobs that call for the skills explored in this talk.

AI Systems and MLOps Engineer for Earth Observation
Forschungszentrum Jülich GmbH
Jülich, Germany
Intermediate
Senior
Linux
Docker
AI Frameworks
Machine Learning

Machine Learning (ML) Engineer Expert - frameworks MLOps / Python / Orchestration/Pipelines
ASFOTEC
Canton de Lille-6, France
Senior
GIT
Bash
DevOps
Python
Gitlab
+6

MLOps Engineer (Kubernetes, Cloud, ML Workflows)
FitNext Co
Charing Cross, United Kingdom
Remote
Intermediate
DevOps
Python
Docker
Grafana
+6


AI & MLOps Engineer - SaaS / AI-Driven Services
Nyou
Linz, Austria
€50-75K
Azure
Python
Kubernetes
Machine Learning
+1

Machine Learning Engineer - Large Language Models (LLM) - Startup
Startup
Charing Cross, United Kingdom
PyTorch
Machine Learning

MLOps / DevOps Engineer (AI/ML & GenAI) Ubicación: España
Talent Connect
Municipality of Madrid, Spain
Bash
Azure
DevOps
Python
Docker
+9

Machine Learning Ops (MLOps) Engineer
Spait Infotech Private Limited
Sheffield, United Kingdom
Remote
£55-120K
Intermediate
ETL
Azure
Scrum
+12

AI Software Engineer - Big Data Pipelines & ML Automation | Python, C#, C++ Expert | Machine Learning Engineer in Manufacturing
Imnoo
Opfikon, Switzerland
Remote
Senior
C++
ETL
.NET
REST
+26