Ron Dagdag
Making neural networks portable with ONNX
#1about 6 minutes
Understanding ONNX as a portable format for ML models
Machine learning models are made portable across different frameworks and hardware using the ONNX open standard, similar to how PDF works for documents.
#2about 2 minutes
When to use ONNX for your machine learning projects
ONNX is ideal for deploying models across different programming languages, achieving low-latency inferencing, and running on resource-constrained edge or IoT devices.
#3about 12 minutes
Four methods for creating or acquiring ONNX models
Models can be obtained from the ONNX Model Zoo, built with tools like Azure Custom Vision, converted from existing frameworks like PyTorch, or used as an intermediary format.
#4about 7 minutes
Deploying models with the high-performance ONNX Runtime
The ONNX Runtime is a high-performance inference engine for deploying models to the cloud or edge devices, bridging the gap between data science and production software engineering.
#5about 4 minutes
Running an ONNX model in a Node.js application
A practical demonstration shows how to load an ONNX model and perform inference within a server-side Node.js application using the `onnxruntime-node` package.
#6about 9 minutes
Performing inference in the browser with ONNX Runtime Web
An emotion detection model is run directly in the browser using ONNX Runtime Web, showcasing client-side inference with JavaScript for privacy and offline capability.
#7about 3 minutes
Optimizing ONNX models for mobile and React Native
ONNX Runtime Mobile provides a lightweight solution for iOS and Android by converting models to a pre-optimized `.ort` format for smaller binary sizes.
#8about 8 minutes
Q&A on starting a career in machine learning
Advice is given on how software developers can enter the machine learning field by starting with model integration and deployment before diving deep into model creation.
Related jobs
Jobs that call for the skills explored in this talk.
Full Stack Developer (all genders welcome)
ROSEN Technology and Research Center GmbH
Osnabrück, Germany
Senior
Matching moments
44:19 MIN
Understanding the ONNX format for model interoperability
Introduction to Azure Machine Learning
23:20 MIN
Deploying machine learning models in a JavaScript environment
Shoot for the moon - machine learning for automated online ad detection
38:21 MIN
Consuming an ONNX model in a .NET console application
Machine Learning in ML.NET
30:21 MIN
Using ONNX Runtime for lightweight model inference
Serverless deployment of (large) NLP models
18:23 MIN
Simplifying development with high-level AI frameworks
Privacy-first in-browser Generative AI web apps: offline-ready, future-proof, standards-based
16:29 MIN
Deploying ML models as a web service with Ktor
The best of both worlds: Combining Python and Kotlin for Machine Learning
14:21 MIN
Comparing platform-dependent and independent model deployment strategies
Developing an AI.SDK
16:28 MIN
Deploying the trained model as an inference service
From Code to Motion: Building an Autonomous Hat-Hunting Robot with Kubernetes & ML
Featured Partners
Related Videos
Privacy-first in-browser Generative AI web apps: offline-ready, future-proof, standards-based
Maxim Salnikov
Generative AI power on the web: making web apps smarter with WebGPU and WebNN
Christian Liebel
Prompt API & WebNN: The AI Revolution Right in Your Browser
Christian Liebel
From ML to LLM: On-device AI in the Browser
Nico Martin
Trends, Challenges and Best Practices for AI at the Edge
Ekaterina Sirazitdinova
How AI Models Get Smarter
Ankit Patel
Your Next AI Needs 10,000 GPUs. Now What?
Anshul Jindal & Martin Piercy
WWC24 - Ankit Patel - Unlocking the Future Breakthrough Application Performance and Capabilities with NVIDIA
Ankit Patel
Related Articles
View all articles


.gif?w=240&auto=compress,format)
From learning to earning
Jobs that call for the skills explored in this talk.

AI Systems and MLOps Engineer for Earth Observation
Forschungszentrum Jülich GmbH
Jülich, Germany
Intermediate
Senior
Linux
Docker
AI Frameworks
Machine Learning


ML Application Engineer (Korean-speaking)
Neural Concept
Lausanne, Switzerland
Fluid
Python
Machine Learning

Web Developer (short-term, 2 months) In Open-Source Machine Learning
Eindhoven University of Technology
Eindhoven, Netherlands
Remote
React
Plotly
Next.js
Machine Learning

ML Application Engineer (German-speaking)
Neural Concept
Großmehring, Germany
Fluid
Python
Machine Learning

ML Application Engineer (French-speaking)
Neural Concept
Lausanne, Switzerland
Fluid
Python
Machine Learning

AI & Embedded ML Engineer (Real-Time Edge Optimization)
autonomous-teaming
Canton of Toulouse-5, France
Remote
C++
GIT
Linux
Python
+1

AI & Embedded ML Engineer (Real-Time Edge Optimization)
autonomous-teaming
München, Germany
Remote
C++
GIT
Linux
Python
+1

AI Technology Student Assignment: Production Deployment of Neural Network Architecture at Signify
Signify International B.V.
Eindhoven, Netherlands
TensorFlow