Applied research project with a UI that dynamically adapts between near-field hand gestures and far-field body gestures based on operator distance. Informed by Fitts' Law analysis and operator workflow observation.
Overview
ExtruFace was an applied research project developing the concept and prototype of a gesture-controlled user interface for industrial extrusion machines. I was involved in all project phases: from requirements analysis through implementation to final acceptance.
The Challenge
Extrusion machines are multi-meter-long, heavy industrial machines for pipe forming. Operators sometimes need to control them from a distance and often wear protective gloves, making the existing touchscreen solution impractical. Key challenges included:
- Adapting modern UX/UI principles to an industrial machinery context
- Designing for two fundamentally different interaction modes: near-field hand gestures and far-field body gestures
- Balancing gesture innovation with operator familiarity and safety requirements
- Ensuring reliability in production environments
The Process
-
Requirements Analysis: Interviews with machine operators to understand workflows and identify which controls are needed at various distances
-
Gesture Design: Findings from my bachelor's thesis significantly influenced the gesture vocabulary; far-mode gestures were designed in collaboration with machine operators
-
Low-Fidelity Prototyping: Wireframes and flowcharts mapping the full interaction flow between near and far mode
-
High-Fidelity Prototype: Functional web application demonstrating both interaction modes
-
User Evaluation: Prototype testing in a laboratory setting with real operators
The Solution
Key features designed and built:
-
Adaptive Interface: UI elements scale dynamically as the operator moves closer to or further from the monitor (via Microsoft Kinect distance tracking)
-
Far Mode: Only distance-relevant controls are displayed, selection based on operator interviews; controlled via body gestures (arm poses) recognized by Microsoft Kinect
-
Near Mode: Touchless hand gesture control using Leap Motion Controller for fine-grained interaction
-
Fitts' Law-Optimized Menus: Circular menus minimizing the distance to menu options for gesture input
-
Touchscreen Fallback: Traditional touch control available as an alternative input method
Results & Impact
- Delivered a functional prototype used as a source of ideas for further development of the existing system
- Demonstrated that touchless gesture control is viable for industrial machine operation
- The adaptive near/far mode concept addressed the core problem of operating machinery from varying distances with protective equipment
Interested in working together?
Get in touch