
International Journal For Multidisciplinary Research
E-ISSN: 2582-2160
•
Impact Factor: 9.24
A Widely Indexed Open Access Peer Reviewed Multidisciplinary Bi-monthly Scholarly International Journal
Home
Research Paper
Submit Research Paper
Publication Guidelines
Publication Charges
Upload Documents
Track Status / Pay Fees / Download Publication Certi.
Editors & Reviewers
View All
Join as a Reviewer
Get Membership Certificate
Current Issue
Publication Archive
Conference
Publishing Conf. with IJFMR
Upcoming Conference(s) ↓
WSMCDD-2025
GSMCDD-2025
AIMAR-2025
Conferences Published ↓
ICCE (2025)
RBS:RH-COVID-19 (2023)
ICMRS'23
PIPRDA-2023
Contact Us
Plagiarism is checked by the leading plagiarism checker
Call for Paper
Volume 7 Issue 4
July-August 2025
Indexing Partners



















Emotion AI:Human Psychology Detection
Author(s) | Dr. Deepali Makarand Bongulwar, Mr. Prathamesh Suryawanshi, Ms. Rajashri Barik, Mr. Prathamesh Shirsat, Ms. Pratiksha Kude |
---|---|
Country | India |
Abstract | This research explores the use of the state-of-the-art transformer-based pre-trained model, the ViT-Face-Expression model from Hugging Face, for human psychology detection from audio and video data. Our technique seeks to provide a full knowledge of human emotions and underlying psychological states by fusing aural cues with visual information gathered from facial expressions. Specifically designed to capture facial emotions within Vision Transformers (ViT), the ViT- Face-Expression model provides a potent tool for deciphering complex emotional nuances stored in facial features. By combining this model with audio processing methods, human psychology can be understood holistically, leading to more accurate identification of emotions and progress in the field of emotion artificial intelligence. Through the integration of visual and aural modalities in emotion recognition, this research advances the rapidly developing subject of Emotion AI. Through the application of audio processing methods in conjunction with the capabilities of the ViT-Face-Expression model, Samlowe model for audio and our goal is to improve the precision and resilience of emotion identification systems. In order to better understand human emotions, our research highlights the significance of taking into account both auditory and visual clues. This will help to develop AI systems that are more sympathetic and context-aware. |
Keywords | Vit Face Expression, Hugging Face, Samlowe, OpenAI, OpenCV |
Field | Engineering |
Published In | Volume 7, Issue 4, July-August 2025 |
Published On | 2025-07-15 |
DOI | https://doi.org/10.36948/ijfmr.2025.v07i04.50599 |
Short DOI | https://doi.org/g9s9mh |
Share this

E-ISSN 2582-2160

CrossRef DOI is assigned to each research paper published in our journal.
IJFMR DOI prefix is
10.36948/ijfmr
Downloads
All research papers published on this website are licensed under Creative Commons Attribution-ShareAlike 4.0 International License, and all rights belong to their respective authors/researchers.
