Skip to content

aminshennan/Hand-Sign-language-detection

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

2 Commits
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

Arabic Sign Language Recognition - Visual Information Processing Project

๐ŸŽฏ Project Overview

This project focuses on developing a comprehensive Arabic Sign Language (ASL) Recognition System using advanced deep learning techniques. Our goal is to bridge communication barriers for the deaf and hard-of-hearing community by creating an accurate and efficient sign language recognition system.

๐Ÿš€ Project Objectives

  • Develop an Arabic Sign Language recognition system using state-of-the-art deep learning models
  • Compare effectiveness of different neural network architectures:
    • Convolutional Neural Networks (CNNs)
    • Long Short-Term Memory (LSTM) Networks
    • Graph Neural Networks (GNNs)
    • Transfer Learning with pre-trained models
  • Evaluate models based on accuracy, efficiency, and practical usability
  • Create a robust data preprocessing pipeline for sign language recognition

๐Ÿ› ๏ธ Technologies & Tools

Deep Learning Frameworks:

  • TensorFlow/Keras - Primary deep learning framework
  • YOLO - Object detection for hand recognition
  • ResNet101V2 - Transfer learning backbone

Libraries & Dependencies:

  • OpenCV - Computer vision and image processing
  • NumPy - Numerical computations
  • Pandas - Data manipulation and analysis
  • Matplotlib - Data visualization and plotting
  • Jupyter Notebook - Development environment

Programming Language:

  • Python - Primary development language

๐Ÿ“ Project Structure

vip/
โ”œโ”€โ”€ README.md                    # Project documentation
โ”œโ”€โ”€ Code/
โ”‚   โ”œโ”€โ”€ code.ipynb              # Main implementation notebook
โ”‚   โ”œโ”€โ”€ best.pt                 # Trained YOLO model (14MB)
โ”‚   โ”œโ”€โ”€ logo.png                # Project logo
โ”‚   โ”œโ”€โ”€ Input/                  # Input data directory
โ”‚   โ””โ”€โ”€ Output/                 # Processed output directory
โ”œโ”€โ”€ Final Report.pdf            # Comprehensive project report
โ””โ”€โ”€ [Presentation File]         # Project presentation

๐Ÿ”ฌ Methodology

1. Data Pipeline

Raw Sign Language Data โ†’ YOLO Hand Detection โ†’ Data Cleaning โ†’ Preprocessing โ†’ Model Training โ†’ Evaluation

2. Data Preprocessing Steps:

  • Hand Detection: YOLO model identifies and extracts hand regions
  • Data Cleaning: Removes irrelevant background and noise
  • Class Organization: Sorts images into class-specific folders
  • Data Augmentation: Enhances dataset diversity

3. Model Development:

  • Exploratory Data Analysis (EDA): Understanding data distribution and characteristics
  • Feature Extraction: Extracting relevant features for sign recognition
  • Model Architecture Design: Implementing various neural network approaches
  • Hyperparameter Tuning: Optimizing model performance

4. Evaluation Metrics:

  • Accuracy - Overall model performance
  • Precision - Correct positive predictions
  • Recall - Ability to find all positive instances
  • F1-Score - Harmonic mean of precision and recall

๐Ÿš€ Getting Started

Prerequisites

Python 3.7+
Jupyter Notebook
CUDA-compatible GPU (recommended)

Installation

  1. Clone the repository:
git clone [repository-url]
cd vip
  1. Install required dependencies:
pip install tensorflow opencv-python numpy pandas matplotlib
pip install torch torchvision  # For YOLO model
pip install jupyter notebook
  1. Navigate to the code directory:
cd Code

Usage

  1. Launch Jupyter Notebook:
jupyter notebook
  1. Open the main implementation:

    • Open code.ipynb in Jupyter Notebook
    • Follow the step-by-step implementation
  2. Data Preparation:

    • Place input sign language images in the Input/ directory
    • Run the data preprocessing cells in the notebook
  3. Model Training:

    • Execute the model training sections
    • Monitor training progress and metrics
  4. Evaluation:

    • Run evaluation cells to assess model performance
    • View results and comparative analysis

๐Ÿ“Š Key Features

๐Ÿค– Advanced Neural Architectures

  • CNN Models: Specialized for image feature extraction
  • LSTM Networks: Capturing temporal dependencies in sign sequences
  • Graph Neural Networks: Modeling hand joint relationships
  • Transfer Learning: Leveraging pre-trained ResNet101V2

๐Ÿ” Intelligent Data Processing

  • YOLO Integration: Accurate hand detection and extraction
  • Automated Cleaning: Removes background noise and irrelevant data
  • Smart Organization: Class-based data structuring

๐Ÿ“ˆ Comprehensive Evaluation

  • Multi-metric Assessment: Accuracy, precision, recall, F1-score
  • Comparative Analysis: Side-by-side model performance evaluation
  • Efficiency Metrics: Training time and inference speed analysis

๐ŸŽฏ Expected Outcomes

  • High-Accuracy Model: Achieving optimal recognition rates for Arabic sign language
  • Efficiency Optimization: Balancing accuracy with computational efficiency
  • Practical Implementation: Ready-to-deploy sign language recognition system
  • Research Contribution: Advancing Arabic sign language recognition research

๐Ÿ“ˆ Results & Performance

Note: Detailed results and performance metrics are available in the Final Report.pdf and within the Jupyter notebook implementation.

Key achievements:

  • โœ… Successful implementation of multiple deep learning architectures
  • โœ… Effective YOLO-based hand detection pipeline
  • โœ… Comprehensive comparative analysis of model performance
  • โœ… Robust data preprocessing and augmentation pipeline

๐Ÿ”ฎ Future Enhancements

  • Real-time Recognition: Implementing live video sign language recognition
  • Mobile Application: Developing mobile app for accessibility
  • Extended Vocabulary: Expanding to larger Arabic sign language vocabulary
  • Multi-language Support: Adding support for other sign languages
  • Edge Deployment: Optimizing models for edge device deployment

๐Ÿ“š Documentation

  • ๐Ÿ“„ Final Report: Comprehensive technical documentation in Final Report.pdf
  • ๐Ÿ’ป Code Documentation: Detailed comments and explanations in code.ipynb
  • ๐ŸŽฏ Methodology: Step-by-step implementation guide in the notebook

๐Ÿค Contributing

This project is part of an academic VIP (Vertically Integrated Projects) program. For collaboration or questions:

  • Academic Institution: Multimedia University (MMU)
  • Project Type: Visual Information Processing

๐Ÿ“ง Contact

For questions or collaboration opportunities, please contact the team members through MMU academic channels.

๐Ÿ™ Acknowledgments

  • Multimedia University (MMU) - Academic support and resources
  • VIP Program - Providing the platform for this research
  • Open Source Community - For the excellent tools and frameworks
  • Arabic Sign Language Community - For the inspiration and importance of this work

About

Arabic sign language recognition system using transfer learning (ResNet101V2) and YOLO-based hand detection. Benchmarks CNN, LSTM, and GNN architectures with multi-metric evaluation framework.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors