Milan Todorovic
Detect Hand Pose with Vision
#1about 2 minutes
Understanding the capabilities of Apple's Vision framework
The Vision framework provides out-of-the-box machine learning tools for image analysis, including object detection, image classification, and tracking faces.
#2about 3 minutes
Recognizing 21 distinct hand landmarks with Vision
Vision processes hand poses by identifying 21 specific landmarks, including the wrist, palm, and four points on each finger and thumb.
#3about 1 minute
Common issues and limitations in hand pose detection
Hand pose recognition can fail due to common real-world issues like partial occlusion, hands near screen edges, wearing gloves, or confusing hands with feet.
#4about 4 minutes
Exploring the structure of the hand tracking Xcode project
The sample application is built around three main components: a CameraView for display, a ViewController for control logic, and a HandGestureProcessor for analyzing gestures.
#5about 2 minutes
Live demo of a drawing app using pinch gestures
A live demonstration shows how to use the tips of the thumb and index finger to create a pinch gesture that draws lines on the iPhone screen.
#6about 4 minutes
Key classes and properties for implementing hand tracking
The implementation relies on key classes like CameraViewController, VNdetectHumanHandPoseRequest for analysis, and UIBezierPath for drawing the visual feedback.
#7about 5 minutes
Processing hand pose observations from the Vision framework
The VNImageRequestHandler processes the camera buffer and returns observations, from which you can extract the coordinates of specific finger joints like the thumb tip.
#8about 2 minutes
Implementing gesture state logic for pinch detection
A custom processor manages gesture states like 'pinched' or 'apart' by calculating the distance between finger landmarks and using a counter for stability.
#9about 1 minute
Applying similar techniques for human body pose detection
The same principles used for hand pose can be applied to full-body pose detection, which tracks major body joints like shoulders, eyes, and ears.
#10about 3 minutes
Exploring potential applications for pose detection
Pose detection technology can be used to build applications that understand sign language, analyze human interaction in images, or create new forms of user input.
Related jobs
Jobs that call for the skills explored in this talk.
Picnic Technologies B.V.
Amsterdam, Netherlands
Intermediate
Senior
Python
Structured Query Language (SQL)
+1
Matching moments
03:16 MIN
Improving the developer feedback loop with specialized tools
Developer Time Is Valuable - Use the Right Tools - Kilian Valkhof
00:59 MIN
Distinguishing real from fake tech headlines
Fake or News: Coding on a Phone, Emotional Support Toasters, ChatGPT Weddings and more - Anselm Hannemann
04:57 MIN
Increasing the value of talk recordings post-event
Cat Herding with Lions and Tigers - Christian Heilmann
03:07 MIN
Final advice for developers adapting to AI
WeAreDevelopers LIVE – AI, Freelancing, Keeping Up with Tech and More
05:17 MIN
Shifting from traditional CVs to skill-based talent management
From Data Keeper to Culture Shaper: The Evolution of HR Across Growth Stages
06:09 MIN
Creating a modal with the native HTML dialog element
WeAreDevelopers LIVE – You Don’t Need JavaScript, Modern CSS and More
00:48 MIN
The shift to on-device AI models in smartphones
Fake or News: Coding on a Phone, Emotional Support Toasters, ChatGPT Weddings and more - Anselm Hannemann
03:58 MIN
Making accessibility tooling actionable and encouraging
Developer Time Is Valuable - Use the Right Tools - Kilian Valkhof
Featured Partners
Related Videos
Let your iOS app read texts
Milan Todorovic
Harnessing Apple Intelligence: Live Coding with Swift for iOS
MIlan Todorović
Apple Vision Pro: Proven Development Methods Meet the Latest Technology
Mario Petricevic
Vision for Websites: Training Your Frontend to See
Daniel Madalitso Phiri
Computer Vision from the Edge to the Cloud done easy
Flo Pachinger
From ML to LLM: On-device AI in the Browser
Nico Martin
Hands-on React Native: From Zero to Hero
Dmitry Vinnik
AR Kit intro - placing 3D objects in a scene and interacting with them in real-time
Nermin Sehic
Related Articles
View all articles



From learning to earning
Jobs that call for the skills explored in this talk.

Apple
Zürich, Switzerland
Python
Pandas
PyTorch
Data analysis
Computer Vision
+1

Apple
Zürich, Switzerland
Python
Pandas
PyTorch
Data analysis
Computer Vision
+1


Plain Concepts
Remote
Azure
Python
Computer Vision
Machine Learning
+2

Vicomtech
Municipality of Bilbao, Spain
Keras
Python
PyTorch
TensorFlow
Data analysis
+3


Paris-based
Paris, France
Python
Docker
TensorFlow
Kubernetes
Computer Vision
+2


Qualitest Group
Zürich, Switzerland
QT
C++
Linux
NumPy
SciPy
+6