Project

Real-Time AR
Gesture-Controlled
Menu System

A touchless human-computer interaction system combining computer vision, AI, and UI design — inspired by Jarvis. Real-time hand tracking with gesture-based menu navigation.

This project builds a real-time AR interface that allows users to interact with a digital menu system using hand gestures instead of traditional input devices. The system uses a webcam to detect and track hand movements, recognize gestures, and control a dynamic on-screen menu.

Objectives

System Architecture

Camera Input ↓ Hand Tracking Module ↓ Gesture Recognition Engine ↓ State Manager (Menu Logic) ↓ AR UI Renderer ↓ User Feedback (Visual Response)

Technologies

Core

  • Python
  • OpenCV (image processing & rendering)
  • MediaPipe (hand tracking & landmarks)

Optional

  • NumPy (math operations)
  • PyAutoGUI (system control)
  • Streamlit (dashboard/debug UI)

Development Phases

Phase Goal Output
1 Hand Tracking Foundation Real-time 21-point landmark visualization
2 Gesture Recognition Engine Swipe, pinch, and hand state detection
3 Menu State System Logical menu structure with state management
4 AR UI Rendering Jarvis-style floating menu overlay
5 Interaction Mapping Gesture-to-action binding
6 Smoothness & UX Cooldowns, smoothing, jitter reduction
7 Feature Expansion Virtual mouse, media control, app launcher
8 Modular Architecture Production-ready codebase

Gesture Mapping

Gesture Action
Swipe Right Next Menu
Swipe Left Previous Menu
Pinch Select Option

Project Timeline

Week 1 OpenCV + Hand Tracking
Week 2 Gesture Detection
Week 3 UI + Menu System
Week 4 Integration + Testing
Week 5 Optimization + Advanced Features

Challenges & Solutions

Challenge Solution
Poor hand detection Improve lighting
Gesture misfires Add thresholds & smoothing
Lag Optimize processing
UI flickering Stabilize coordinates

Real-World Applications

Project Structure

project/ ├── hand_tracking.py ├── gesture_engine.py ├── ui_renderer.py ├── main.py