GestureGenius transforms your standard webcam into a touchless virtual cursor controller with high precision and real-time responsiveness. This project leverages powerful computer vision libraries—OpenCV and Mediapipe—to track hand gestures and seamlessly translate them into on-screen actions like clicks, movements, and scrolling via PyAutoGUI.
Designed for accessibility, gaming, smart environments, and interactive displays, GestureGenius pushes the boundaries of modern human-computer interaction, offering a futuristic and intuitive alternative to traditional input devices.
- Features
- Motivation
- Installation
- Usage
- Methodology
- Applications
- Future Plans
- Contributions
- License
- Literature Survey & References
- Hand Gesture Recognition: Detect hand movements via webcam.
- Real-time Cursor Control: Use gestures to control the cursor and perform actions like clicking and scrolling.
- No Specialized Hardware Needed: Works with standard webcams.
- Cross-Platform Compatibility: Supports Windows, Linux, and macOS.
- Seamless Integration: Designed for accessibility, gaming, presentations, and more.
Traditional input devices—like keyboards and mice—create a barrier between users and digital content. Gesture-based interfaces bring a natural, immersive experience, especially beneficial for users with accessibility needs, or in gaming, smart displays, and automotive control systems. GestureGenius provides a smooth, touch-free solution to empower modern interactions.
Ensure that Python 3.x is installed. You can get it from Python's official website.
Clone this repository to your local machine:
git clone https://github.com/achyuth-2308/GestureGenius-Virtual-Cursor-Commander.git
cd GestureGenius-Virtual-Cursor-CommanderInstall the required Python libraries by running:
pip install -r requirements.txtOr install them individually:
pip install opencv-python mediapipe pyautogui- Webcam: Any standard webcam (built-in or external).
- Processor: Intel i3 or higher (i5 or above recommended).
- RAM: 4GB or more.
- Operating System: Windows, Linux, or macOS.
Ensure your webcam is connected and operational.
Open your terminal and navigate to the project folder. Run the following command:
python gesturegenius.pyOnce the webcam feed is live:
- Move Cursor: Move your hand to control the cursor.
- Click: Pinch your fingers together for a mouse click.
- Scroll: Swipe your hand up or down to scroll.
Press Ctrl + C in your terminal to stop the program.
GestureGenius relies on a sophisticated pipeline:
- Hand Detection: OpenCV processes the video input from your webcam and detects hand regions.
- Landmark Extraction: Mediapipe identifies 21 key hand landmarks, allowing precise tracking.
- Gesture Recognition: Custom algorithms interpret gestures (e.g., pinch for clicks, swipes for scrolling).
- Cursor Mapping: PyAutoGUI converts gestures into system-wide actions, such as mouse movement and scrolling.
- Accessibility: Assist individuals with disabilities to control their computer through gestures.
- Interactive Displays: Enable touch-free interactions in kiosks and public systems.
- Gaming: Adds an immersive element to gameplay by controlling the environment using gestures.
- Presentations: Navigate slides with hand gestures during business or academic presentations.
- Automotive Interfaces: Enable gesture controls for infotainment systems in vehicles.
- Smart Home Systems: Control smart home displays or devices without touching a screen.
- Expand Gesture Set: Add more complex gestures like multi-finger movements or 3D gestures.
- Improve Accuracy: Integrate machine learning models to further improve detection and reduce false positives.
- Cross-Application Support: Expand support to game engines, automotive interfaces, and smart home systems.
We’re excited to collaborate! Feel free to open issues, submit pull requests, or suggest new features. Let's push the boundaries of human-computer interaction together.
This project is licensed under the MIT License. See the LICENSE file for details.
Gesture recognition and computer vision have become pivotal in enhancing human-computer interaction. Libraries like OpenCV and Mediapipe have democratized access to these technologies, making gesture-based interfaces feasible for everyday users without requiring specialized hardware.
- Zhang, Z. (2021). Hand Tracking in Computer Vision Using Mediapipe. International Journal of Computer Vision.
- Bradski, G. (2000). The OpenCV Library. Dr. Dobb’s Journal of Software Tools.
- Raffel, C. (2020). Gesture Recognition: A Comprehensive Survey. IEEE Transactions on Human-Machine Systems.
- Gutierrez, M. (2022). Human-Computer Interaction through Vision-Based Systems. ACM Computing Surveys.
This repository hosts the code and documentation for GestureGenius: Virtual Cursor Commander. Contributions and feedback are highly encouraged as we push the boundaries of hands-free, gesture-based interaction for a smarter, touchless future!