Skip to content

WixBe/Emoshott

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

EMO-SHOTT : AI Emotional Trigger Detector

Emotional well-being is crucial for a healthy life, impacting physical health, mental health, and relationships. Understanding and managing emotions can enhance overall well-being. Our project aims to bridge the gap in current systems by detecting emotional triggers through multimodal analysis of facial expressions and speech.

Key Features

  • Emotion Trigger Detection: Combines facial motion analysis and speech-to-text processing.
  • Enhanced Emotion Classification: Expands recognition from 3 to 8 emotions – Anger, Disgust, Fear, Happiness, Sadness, Surprise, Neutral, and Contempt.
  • Data Processing:
    • FER-2013 PLUS Dataset: 48x48 pixel grayscale face images with improved labeling.
    • OpenAI Whisper: Speech recognition and translation for audio transcriptions.
    • spaCy NLP: Part-of-Speech tagging and Named Entity Recognition (NER) for trigger identification.

Methodology

Picture1

Facial Analysis

  • Feature Extraction using Haar features.
  • AdaBoost for feature selection and classification.
  • Cascade Classifiers for improved efficiency.

Speech Processing

  • Audio split into 30-sec chunks, processed via Whisper encoder-decoder.
  • NLP mapping of triggers using spaCy.

Model Training

  • Tested various architectures including VGG16, TensorFlow Keras, PyTorch DDAMNET, and CNN.
  • Achieved validation accuracy of 86.36% after 100 epochs. Picture2 Picture3 Picture4

Achievements

  • Successfully mapped emotions to triggers by correlating subtitles and apex frames.
  • Developed a robust system enhancing mental well-being through accurate emotion detection.

About

An AI emotional trigger identifier

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages