Skip to content

Latest commit

 

History

History
108 lines (66 loc) · 5.51 KB

README.md

File metadata and controls

108 lines (66 loc) · 5.51 KB

GestureFlow: A Novel Hand Gesture Control System for Interactive Gaming

Research Paper Link

Overview

Welcome to the GestureFlow project! This project aims to revolutionize interactive gaming by introducing a novel hand gesture control system. By leveraging advanced computer vision techniques, we have developed a system that allows players to control games using natural hand movements, providing a more immersive and engaging gaming experience.

Table of Contents

  1. Introduction
  2. Literature Survey
  3. Methodology
  4. Results and Discussion
  5. Conclusion
  6. References

Introduction

The gaming industry is at a pivotal point where alternative input methods are being explored to enhance user interaction. GestureFlow aims to improve the immersive gaming experience by providing interfaces that capture hand gestures to control actions within games. This approach integrates computer vision, hand segmentation, and gesture recognition to offer a natural and engaging way to interact with virtual environments.

Literature Survey

Our research builds upon several key studies in the field of hand gesture recognition and interactive gaming. Notable contributions include:

  • Multi-modal zero-shot dynamic hand gesture recognition using Transformer and other advanced models.
  • Real-time hand gesture to text translation with high accuracy.
  • Gesture-controlled virtual artboards for communication and education.

Methodology

Camera Image Acquisition

We use OpenCV, DirectShow (Windows), and Video4Linux2 (V4L2) for capturing images and videos from cameras. These tools provide a robust framework for real-time image and video processing.

Background and Hand Segmentation

Two primary techniques are used:

  • Skin Colour Detection: Identifies hand pixels based on skin color.
  • Background Subtraction: Uses a stable background to isolate hand movements.

Hand Detection Framework

Google’s Mediapipe Framework is employed for real-time hand detection using feature points of hands. It provides accurate 2D and 3D hand landmark coordinates for gesture recognition.

Hand Detection Mechanism

  • Landmark Identification: Detects key points on the hand such as the index finger tip and wrist.
  • Angle Calculation: Calculates angles formed by the fingers to determine gestures.

Development of Game

Using libraries like Mediapipe, OpenCV, and Pygame, we developed a game where gestures control the game actions. Pygame handles game events, logic, and graphics.

Game Environment and Logic

The game environment is initialized with assets like car images and background. The main game loop handles player input, updates game state, and displays the game.

User Interface and Management

Messages like “Game Over” are displayed based on game events. The background is continuously updated based on game speed and screen dimensions.

Mapping Function

A custom mapping function associates each recognized gesture with a specific control command. This function uses data structures like hash maps and arrays for efficient gesture-action mappings.

Gesture-Based Directional Control

The system monitors hand movements and maps them to game controls. The car game can also be controlled using a keyboard as an alternative input method.

Mapping Gestures to Control

The mapping method improves user experience and the intuitiveness of gesture-based interactions within the game.

Optimization and Performance Considerations

Optimization techniques like parallel processing and hardware acceleration minimize latency and enhance system performance.

Validation and User Feedback

Performance tests and user feedback loops are used to improve the gesture-action mapping scheme based on gameplay scenarios and user interactions.

Results and Discussion

Preliminary results show that the gesture-based control method is effective and engaging. Users found the system natural and easy to use, though some lags and errors were noted. Future improvements will focus on enhancing system response and accuracy.

Conclusion

GestureFlow represents a significant advancement in integrating hand gesture control systems with interactive gaming. By providing a comprehensive understanding of the technical intricacies involved, this research contributes to the advancement of gesture-based interaction paradigms in gaming.

References

For a detailed list of references and further reading, please refer to the full research paper included in this repository.


We hope you find GestureFlow to be an exciting and innovative approach to interactive gaming. Thank you for exploring our project!