International Journal of Emerging Research in Science, Engineering, and Management
Vol. 2, Issue 4, pp. 64-72, April 2026.
This work is licensed under a Creative Commons Attribution 4.0 International License.
AI Simulator that Quantifies interview Performance with Real-Time Analytics
E Murali
M Rekha
Kuppam Nithin
Sirisha Pathapati
Madduru Sucharitha
Varalekhari Neeraj Kumar
1Professor, Department of CSE, Siddharth Institute of Engineering & Technology, Puttur, AP, India.
2Assistant Professor, Department of CSE, Siddharth Institute of Engineering & Technology, Puttur, AP, India.
3-5UG Scholar, Department of CSE, Siddharth Institute of Engineering & Technology, Puttur, AP, India.
Abstract: Traditional interview preparation often fails to ad- dress the critical role of non-verbal communication, leaving candidates unaware of how their confidence and emotional state are perceived. To bridge this gap, this method presents an AI-driven mock interview simulator that provides a holistic evaluation of a candidate’s performance. The system leverages a multi- modal AI pipeline to analyze a confluence of verbal and non- verbal cues. Key modalities include facial expression analysis for emotional sentiment, vocal tonality assessment for confidence markers such as pitch and pace, and linguistic pattern analysis of the transcribed speech to measure clarity and identify filler word usage. The platform delivers a tailored experience by generating personalized interview questions and culminates in a comprehensive performance report. This report presents users with quantitative metrics and actionable insights, pinpointing specific emotional cues and highlighting concrete areas for development. By providing a safe, data-driven environment for practice, this tool empowers candidates to master the crucial soft skills required to succeed, thereby enhancing their confidence and competitiveness in the modern job market.
Keywords: Multi-model AI Pipeline, Mock Interview Simulator, Real-Time Feedback, Facial Emotion Recognition, Vocal Tone Analysis.
References:
- G H. Touvron et al., “LLAMA: Open and Efficient Foundation Language Models,” arXiv (Cornell University), Feb. 2023, doi: 10.48550/arxiv.2302.13971.
- S. I. Serengil and A. Ozpinar, “LightFace: A Hybrid Deep Face Recognition Framework,” 2020 Innovations in Intelligent Systems and Applications Conference (ASYU), Istanbul, Turkey, 2020, pp. 1-5, doi: 10.1109/ASYU50717.2020.9259802.
- B. McFee et al., “librosa: Audio and Music Signal Analysis in Python,” Proceedings of the Python in Science Conferences, pp. 18–24, Jan. 2015, doi: 10.25080/majora-7b98e3ed-003.
- M. Nalluri, M. Pentela, and N. R. Eluri, “A scalable tree boosting system: XG Boost,” ResearchGate, Jan. 2020, doi: 10.22259/2349-476X.0712005.
- C. Lugaresi et al., “MediaPipe: A framework for building perception Pipelines,” arXiv (Cornell University), Jun. 2019, doi: 10.48550/arxiv.1906.08172.
- N. Reimers and I. Gurevych, “Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks,” TUbilio (Technical University of Darmstadt), Aug. 2019, doi: 10.48550/arxiv.1908.10084.
- J. Kim, J. Kong, and J. Son, “Conditional Variational Autoencoder with Adversarial Learning for End-to-End Text-to-Speech,” arXiv (Cornell University), Jun. 2021, doi: 10.48550/arxiv.2106.06103.
- L. Chen et al., “Automated scoring of nonnative speech using the SpeechRaterSMV. 5.0 engine,” ETS Research Report Series, vol. 2018, no. 1, pp. 1–31, Apr. 2018, doi: 10.1002/ets2.12198.
- S. Poria, E. Cambria, R. Bajpai, and A. Hussain, “A review of affective computing: From unimodal analysis to multimodal fusion,” Information Fusion, vol. 37, pp. 98–125, Feb. 2017, doi: 10.1016/j.inffus.2017.02.003.
- Y. Adepu, V. R. Boga and S. U, “Interviewee Performance Analyzer Using Facial Emotion Recognition and Speech Fluency Recognition,” 2020 IEEE International Conference for Innovation in Technology (INOCON), Bangluru, India, 2020, pp. 1-5, doi: 10.1109/INOCON50539.2020.9298427.
- G. S. Rao, M. Jaiganesh, M. A. K, M. A. A. A, N. R and P. K. Parida, “AI Powered Virtual Job Interview Simulator Using Natural Language Processing,” 2025 8th International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 2025, pp. 1413-1421, doi: 10.1109/ICOEI65986.2025.11013504.
- S. R. Jagtap, “AI-Driven Real- Time Interview Simulation App with Voice Recognition and Facial Analysis,” Indian Journal of Science and Technology, vol. 18, no. 25, pp. 2058–2066, Jun. 2025, doi: 10.17485/ijst/v18i25.760.
