Ava-Artificial Virtual Assistant
DOI:
https://doi.org/10.62647/IJITCE2025V13I3PP40-45Keywords:
Artificial Intelligence, virtual assistant, multimodal interactions, Natural Language Processing (NLP), speech recognition, YOLOv8, object detection, OpenCV, AI processing, deep learning.Abstract
AVA (Artificial Virtual Assistant) is a smart, multimodal AI system that enables interaction through voice, text, and images, offering a natural and dynamic user experience. Unlike traditional assistants, AVA supports real-time web access, on device AI processing, object detection, and image-to-text conversion, making it capable of analyzing and responding to both visual and spoken content intelligently. Developed using Python and Kivy, it incorporates technologies like YOLOv8[1] for object detection, speech recognition for voice interaction, and MongoDB for local data storage, ensuring strong privacy and performance by minimizing reliance on cloud services. AVA includes a rich set of features such as email automation, WhatsApp messaging, Google and Wikipedia search, YouTube video playback, weather updates, language translation, and the ability to launch desktop applications like Notepad and Camera. Its modular design and user-friendly interface make it suitable for various real-world applications, and its scalability allows for future enhancements such as emotion detection, gesture control, and multilingual support, establishing AVA as a powerful and adaptable virtual assistant[2].
Downloads
Downloads
Published
Issue
Section
License
Copyright (c) 2025 Authors

This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.