Welcome To The Voca Guide
AI-Powered Spoken Language-Learning Mobile Application for visually impaired children in Sri Lanka
Read MoreAI-Powered Spoken Language-Learning Mobile Application for visually impaired children in Sri Lanka
Read MoreVocaGuide is more than just a language-learning app. It’s an AI-powered companion designed especially for visually impaired children in Sri Lanka. With its conversational AI, it communicates naturally with children, making learning feel like a friendly dialogue rather than a lesson. The app intelligently detects language weaknesses and uses engaging quiz-based gamification to target those areas, ensuring steady improvement in vocabulary, speaking, and comprehension. By blending accessibility, personalization, and fun, VocaGuide creates an inclusive learning experience that empowers every child to build confidence in language skills.
A visually impaired child receives quizzes generated based on their
identified language weaknesses and current skill level. The questions are
presented interactively using text-to-speech and speech-to-text, allowing the child to
respond verbally. This approach makes practicing language engaging and accessible,
while targeting areas that need improvement for steady progress.
A visually impaired child can start conversations directly with the app, which responds naturally using
conversational AI. This creates a safe and interactive environment for practicing spoken
language, helping the child improve pronunciation, tone, and fluency while building
confidence in everyday communication.
While the child engages in conversation with the app, it performs phoneme-level evaluation of their speech. The system identifies the user’s spoken language proficiency and tracks progress over time, providing insights that help guide further practice and improvement.
The app integrates a custom-trained intelligent voice assistant to provide audio guidance, creating a smooth and adaptive user experience. Its design prioritizes accessibility, ensuring that visually impaired children can navigate, interact, and learn independently with ease.
Read our research domain details here
VocaGuide is an AI-powered spoken language learning app designed for visually impaired children in Sri Lanka. It combines conversational AI, quiz-based practice, and speech assessment to create a personalized, interactive learning experience.
The app identifies language weaknesses and generates quizzes tailored to each child’s level, delivered via text-to-speech and speech-to-text for engaging practice. Conversational AI enables natural dialogue, helping children improve pronunciation, tone, and fluency. Phoneme-level speech assessment tracks progress over time, while accessible UI and a custom voice assistant ensure a smooth, adaptive learning environment.
VocaGuide makes language education inclusive, interactive, and empowering for every child.
Read our research domain details here
The integration of artificial intelligence (AI) into mobile applications has opened new opportunities for inclusive education, particularly for visually impaired children in Sri Lanka. Traditional language-teaching methods often rely heavily on printed materials and classroom-based approaches, which limit accessibility for this group. As a result, there is a growing need for innovative solutions that combine technology with accessibility to create more engaging and adaptive learning environments. AI-powered spoken language-learning tools such as VocaGuide represent a promising step toward addressing this gap.
One of the key areas where such applications provide value is in delivering personalized learning experiences. By identifying each child’s language weaknesses and generating targeted practice activities, these tools offer a level of adaptability that traditional methods cannot match. Features like quiz-based gamification further enhance motivation by transforming learning into an interactive experience, while conversational AI creates a safe environment where children can practice spoken language without fear of judgment. These approaches align with global trends in adaptive education, where technology is increasingly used to tailor content to individual needs.
Another critical consideration is accessibility and ease of use. For visually impaired learners, audio guidance and speech technologies ensure that learning can be both independent and engaging. At the same time, integrating culturally and linguistically relevant content-particularly in Sinhala and Tamil-ensures that the solution remains contextually meaningful for Sri Lankan children. Despite global advances in assistive learning, Sri Lanka continues to face challenges in providing adequate resources for children with disabilities. Applications like VocaGuide have the potential to bridge this gap by leveraging AI and mobile technologies to deliver cost-effective, scalable, and impactful learning solutions for an underserved community.
There is a clear research gap in the development of AI-powered spoken language-learning applications for visually impaired children in Sri Lanka. While global solutions in speech recognition and grammar correction have advanced rapidly, very few are tailored to the unique linguistic, cultural, and accessibility needs of Sri Lankan learners. Current approaches often fail to capture the phoneme-level detail and grammar variation present in Sri Lankan English, and existing gamification methods remain heavily visual, leaving blind learners at a disadvantage. To bridge these gaps, dedicated solutions like VocaGuide are essential, ensuring that accessibility is not treated as an afterthought but as the foundation of design.
1. Audio-Only Gamification for Visually Impaired Learners
* Many language-learning apps focus on visual gamification, making them unsuitable for visually impaired children.
* Current tools rely on static question banks and generic feedback, with little adaptation to individual learner errors.
* There is a lack of real-time, speech-driven micro-quizzes and spaced repetition delivered through natural, child-friendly TTS.
2. Mispronunciation & Grammar Error Detection
* Existing systems report sentence-level accuracy but lack phoneme-level diagnostics that highlight segmental and suprasegmental issues common in Sri Lankan English.
* Current classroom products rely on visual feedback (charts, heatmaps) rather than concise audio explanations accessible to blind learners.
* Grammar correction tools tend to over-correct Sri Lankan English forms, lowering trust and usability.
3. Conversational Grammar Error Correction
* Most grammar correction systems ignore conversational context, cultural nuance, and pragmatics, leading to rigid or irrelevant feedback.
* Existing models struggle with efficiency on mobile devices, often suffering from latency and high memory demands.
* Accessibility gaps remain, as few systems are designed to be speech-first, low-latency, and inclusive of multiple disabilities.
The main research problem addressed in this study is the lack of accessible, audio-first language learning tools for visually impaired children in Sri Lanka. Existing mobile applications for language learning are predominantly visual, making them unsuitable for children who are blind or have low vision. As a result, these learners have very limited opportunities to practice spoken English in an engaging and accessible environment.
Current speech recognition systems often fail to capture the unique features of Sri Lankan English pronunciation, overlooking common segmental and suprasegmental errors that affect communication. At the same time, generic grammar correction models frequently misidentify valid Sri Lankan English usage as incorrect, which can discourage learners and undermine trust in the system. Most available solutions also depend on reliable internet connectivity and high-end devices, which makes them inaccessible in low-resource educational settings. Together, these issues leave visually impaired learners without credible, personalized support for improving their spoken language abilities.
Our research aims to design and evaluate an inclusive, speech-first tutoring system that provides accent-sensitive pronunciation feedback, delivers trustworthy grammar suggestions adapted to Sri Lankan English, and incorporates audio-based gamification to keep learners motivated. The solution will operate effectively on mid-range mobile devices, function offline when needed, and prioritize user privacy. By addressing these challenges, the project seeks to create a practical and scalable tool that improves language skills while ensuring accessibility for visually impaired children in Sri Lanka.
The main objective of this research is to design, develop, and evaluate an audio-only, conversational, and gamified English language learning system tailored for visually impaired Sri Lankan learners. The system aims to provide accent-aware pronunciation support, context-sensitive grammar correction, and interactive quiz-based practice, all delivered through fully accessible, speech-first interfaces. By focusing on accessibility, cultural relevance, and privacy, the solution seeks to empower learners to practice independently, build confidence in spoken English, and achieve measurable educational outcomes under real-world resource constraints such as limited connectivity and mid-range devices.
1. Design an adaptive quiz-based gamification engine that generates voice-only practice activities from detected weaknesses, with spaced repetition for sustained progress.
2. Implement a conversational AI assistant that supports multi-turn spoken interaction, role-play scenarios, and peer-style dialogue to encourage natural language use.
3. Develop a phoneme-level pronunciation assessment module that detects and explains mispronunciations in Sri Lankan English with accessible audio feedback.
The methodology for developing Vocaguide involves a mixed-methods approach. First, we will conduct qualitative research through interviews and focus groups with visually impaired learners and educators to identify challenges in spoken language learning. Second, we will carry out a quantitative phase using surveys and pilot studies to collect data on learner needs, technology use, and baseline performance.
We will then train and fine-tune speech and language models for phoneme recognition, pronunciation scoring, and grammar correction, and implement these in a React Native + FastAPI application designed with accessibility-first principles. Usability testing and iterative prototyping will be conducted to refine the system. Finally, learning outcomes will be evaluated through pre- and post-test comparisons, usage analytics, and learner feedback to ensure effectiveness and continuous improvement.
React Native (Expo)
FastAPI
Python
Jupyter Notebook
Anaconda Navigator
Postman API Platform
VS Code
Figma UI/UX
GitHub
ngrok
Proposal presentation and the proposal report submission.
50% progress presentation of the research project.
90% progress presentation of the research project.
Submission and presentation of the camera-ready research poster.
Submission of final reports and the final presentation of the research.
Final viva of the research Commerclization Video with an User Testing and Research Team Member.
View Our Presentations Here

Head of Department
kalpani.m@sliit.lk
FACULTY OF COMPUTING | COMPUTER SCIENCE & SOFTWARE ENGINEERING

Lecturer
jeewaka.p@sliit.lk
FACULTY OF COMPUTING | COMPUTER SCIENCE & SOFTWARE ENGINEERING

IT21809088
it21809088@my.sliit.lk
FACULTY OF COMPUTING | SOFTWARE ENGINEERING

IT21838002
it21838002@my.sliit.lk
FACULTY OF COMPUTING | SOFTWARE ENGINEERING

IT19211688
it19211688@my.sliit.lk
FACULTY OF COMPUTING | SOFTWARE ENGINEERING
SLIIT Malabe Campus, New Kandy Rd, Malabe, SriLanka
vocaguideofficial@gmail.com
+94 70 2648 320