Welcome To The Voca Guide

AI-Powered Spoken Language-Learning Mobile Application for visually impaired children in Sri Lanka

Read More


Why Choose VocaGuide ?

VocaGuide is more than just a language-learning app. It’s an AI-powered companion designed especially for visually impaired children in Sri Lanka. With its conversational AI, it communicates naturally with children, making learning feel like a friendly dialogue rather than a lesson. The app intelligently detects language weaknesses and uses engaging quiz-based gamification to target those areas, ensuring steady improvement in vocabulary, speaking, and comprehension. By blending accessibility, personalization, and fun, VocaGuide creates an inclusive learning experience that empowers every child to build confidence in language skills.

Quiz-Based Gamification


A visually impaired child receives quizzes generated based on their identified language weaknesses and current skill level. The questions are presented interactively using text-to-speech and speech-to-text, allowing the child to respond verbally. This approach makes practicing language engaging and accessible, while targeting areas that need improvement for steady progress.

Conversational AI


A visually impaired child can start conversations directly with the app, which responds naturally using conversational AI. This creates a safe and interactive environment for practicing spoken language, helping the child improve pronunciation, tone, and fluency while building confidence in everyday communication.

Speech and Pronunciation Assessment

While the child engages in conversation with the app, it performs phoneme-level evaluation of their speech. The system identifies the user’s spoken language proficiency and tracks progress over time, providing insights that help guide further practice and improvement.

UI/UX Design and Accessibility

The app integrates a custom-trained intelligent voice assistant to provide audio guidance, creating a smooth and adaptive user experience. Its design prioritizes accessibility, ensuring that visually impaired children can navigate, interact, and learn independently with ease.

About the Project

Read our research domain details here

Enhancing Language Learning for Visually Impaired Children with VocaGuide.

VocaGuide is an AI-powered spoken language learning app designed for visually impaired children in Sri Lanka. It combines conversational AI, quiz-based practice, and speech assessment to create a personalized, interactive learning experience.

The app identifies language weaknesses and generates quizzes tailored to each child’s level, delivered via text-to-speech and speech-to-text for engaging practice. Conversational AI enables natural dialogue, helping children improve pronunciation, tone, and fluency. Phoneme-level speech assessment tracks progress over time, while accessible UI and a custom voice assistant ensure a smooth, adaptive learning environment.

VocaGuide makes language education inclusive, interactive, and empowering for every child.

Research Project Scope

Read our research domain details here

Literature Survey

The integration of artificial intelligence (AI) into mobile applications has opened new opportunities for inclusive education, particularly for visually impaired children in Sri Lanka. Traditional language-teaching methods often rely heavily on printed materials and classroom-based approaches, which limit accessibility for this group. As a result, there is a growing need for innovative solutions that combine technology with accessibility to create more engaging and adaptive learning environments. AI-powered spoken language-learning tools such as VocaGuide represent a promising step toward addressing this gap.

One of the key areas where such applications provide value is in delivering personalized learning experiences. By identifying each child’s language weaknesses and generating targeted practice activities, these tools offer a level of adaptability that traditional methods cannot match. Features like quiz-based gamification further enhance motivation by transforming learning into an interactive experience, while conversational AI creates a safe environment where children can practice spoken language without fear of judgment. These approaches align with global trends in adaptive education, where technology is increasingly used to tailor content to individual needs.

Another critical consideration is accessibility and ease of use. For visually impaired learners, audio guidance and speech technologies ensure that learning can be both independent and engaging. At the same time, integrating culturally and linguistically relevant content-particularly in Sinhala and Tamil-ensures that the solution remains contextually meaningful for Sri Lankan children. Despite global advances in assistive learning, Sri Lanka continues to face challenges in providing adequate resources for children with disabilities. Applications like VocaGuide have the potential to bridge this gap by leveraging AI and mobile technologies to deliver cost-effective, scalable, and impactful learning solutions for an underserved community.

Research Gap

There is a clear research gap in the development of AI-powered spoken language-learning applications for visually impaired children in Sri Lanka. While global solutions in speech recognition and grammar correction have advanced rapidly, very few are tailored to the unique linguistic, cultural, and accessibility needs of Sri Lankan learners. Current approaches often fail to capture the phoneme-level detail and grammar variation present in Sri Lankan English, and existing gamification methods remain heavily visual, leaving blind learners at a disadvantage. To bridge these gaps, dedicated solutions like VocaGuide are essential, ensuring that accessibility is not treated as an afterthought but as the foundation of design.

1. Audio-Only Gamification for Visually Impaired Learners

* Many language-learning apps focus on visual gamification, making them unsuitable for visually impaired children.

* Current tools rely on static question banks and generic feedback, with little adaptation to individual learner errors.

* There is a lack of real-time, speech-driven micro-quizzes and spaced repetition delivered through natural, child-friendly TTS.

2. Mispronunciation & Grammar Error Detection

* Existing systems report sentence-level accuracy but lack phoneme-level diagnostics that highlight segmental and suprasegmental issues common in Sri Lankan English.

* Current classroom products rely on visual feedback (charts, heatmaps) rather than concise audio explanations accessible to blind learners.

* Grammar correction tools tend to over-correct Sri Lankan English forms, lowering trust and usability.

3. Conversational Grammar Error Correction

* Most grammar correction systems ignore conversational context, cultural nuance, and pragmatics, leading to rigid or irrelevant feedback.

* Existing models struggle with efficiency on mobile devices, often suffering from latency and high memory demands.

* Accessibility gaps remain, as few systems are designed to be speech-first, low-latency, and inclusive of multiple disabilities.

Research Problem

The main research problem addressed in this study is the lack of accessible, audio-first language learning tools for visually impaired children in Sri Lanka. Existing mobile applications for language learning are predominantly visual, making them unsuitable for children who are blind or have low vision. As a result, these learners have very limited opportunities to practice spoken English in an engaging and accessible environment.

Current speech recognition systems often fail to capture the unique features of Sri Lankan English pronunciation, overlooking common segmental and suprasegmental errors that affect communication. At the same time, generic grammar correction models frequently misidentify valid Sri Lankan English usage as incorrect, which can discourage learners and undermine trust in the system. Most available solutions also depend on reliable internet connectivity and high-end devices, which makes them inaccessible in low-resource educational settings. Together, these issues leave visually impaired learners without credible, personalized support for improving their spoken language abilities.

Our research aims to design and evaluate an inclusive, speech-first tutoring system that provides accent-sensitive pronunciation feedback, delivers trustworthy grammar suggestions adapted to Sri Lankan English, and incorporates audio-based gamification to keep learners motivated. The solution will operate effectively on mid-range mobile devices, function offline when needed, and prioritize user privacy. By addressing these challenges, the project seeks to create a practical and scalable tool that improves language skills while ensuring accessibility for visually impaired children in Sri Lanka.

Research Objectives

Main Objective

The main objective of this research is to design, develop, and evaluate an audio-only, conversational, and gamified English language learning system tailored for visually impaired Sri Lankan learners. The system aims to provide accent-aware pronunciation support, context-sensitive grammar correction, and interactive quiz-based practice, all delivered through fully accessible, speech-first interfaces. By focusing on accessibility, cultural relevance, and privacy, the solution seeks to empower learners to practice independently, build confidence in spoken English, and achieve measurable educational outcomes under real-world resource constraints such as limited connectivity and mid-range devices.

Sub Objectives

1. Design an adaptive quiz-based gamification engine that generates voice-only practice activities from detected weaknesses, with spaced repetition for sustained progress.

2. Implement a conversational AI assistant that supports multi-turn spoken interaction, role-play scenarios, and peer-style dialogue to encourage natural language use.

3. Develop a phoneme-level pronunciation assessment module that detects and explains mispronunciations in Sri Lankan English with accessible audio feedback.







Methodology

The methodology for developing Vocaguide involves a mixed-methods approach. First, we will conduct qualitative research through interviews and focus groups with visually impaired learners and educators to identify challenges in spoken language learning. Second, we will carry out a quantitative phase using surveys and pilot studies to collect data on learner needs, technology use, and baseline performance.

We will then train and fine-tune speech and language models for phoneme recognition, pronunciation scoring, and grammar correction, and implement these in a React Native + FastAPI application designed with accessibility-first principles. Usability testing and iterative prototyping will be conducted to refine the system. Finally, learning outcomes will be evaluated through pre- and post-test comparisons, usage analytics, and learner feedback to ensure effectiveness and continuous improvement.

Tools & Technologies

React Native (Expo)

FastAPI

Python

Jupyter Notebook

Anaconda Navigator

Postman API Platform

VS Code

Figma UI/UX

GitHub

ngrok

Milestones

Project Proposal

Proposal presentation and the proposal report submission.


Progress Presentation 1

50% progress presentation of the research project.


Progress Presentation 2

90% progress presentation of the research project.


Demonstration

Submission and presentation of the camera-ready research poster.


Final Assessment

Submission of final reports and the final presentation of the research.


Viva

Final viva of the research Commerclization Video with an User Testing and Research Team Member.


Documentation

Project Charter

Click Here to View the Project Charter Document

R25-071_Project_Charter

Proposal Document

Access the Individual Proposal Documents Using Following Links

IT21809088_Sandeepa K.B.A.R.

IT21838002_Jayawardhana A.M.S.P.

IT19211688_Weerasinghe C.D.

Research Logbook

Access the Research Logbook Documents Using Following Links

IT21809088_Sandeepa K.B.A.R.

IT21838002_Jayawardhana A.M.S.P.

IT19211688_Weerasinghe C.D.

Final Reports

Access the Final Group Report Using Following Link

R25-071_Group_Report

Access the Individual Final Reports Using Following Links

IT21809088_Sandeepa K.B.A.R.

IT21838002_Jayawardhana A.M.S.P.

IT19211688_Weerasinghe C.D.

Presentations

View Our Presentations Here

Proposal Presentation

Click to View the Proposal Presentation

Progress Presentation I

Click to View the Progress Presentation I - 50% Completion

Progress Presentation II

Click to View the Progress Presentation II - 90% Completion

Final Presentation

Click to View the Final Presentation- 100% Completion

Our Team

Dr. Kalpani Manathunga

SUPERVISOR

Head of Department

kalpani.m@sliit.lk

FACULTY OF COMPUTING | COMPUTER SCIENCE & SOFTWARE ENGINEERING

Mr. Jeewaka Perera

Co-SUPERVISOR

Lecturer

jeewaka.p@sliit.lk

FACULTY OF COMPUTING | COMPUTER SCIENCE & SOFTWARE ENGINEERING

Sandeepa K.B.A.R.

TEAM Leader

IT21809088

it21809088@my.sliit.lk

FACULTY OF COMPUTING | SOFTWARE ENGINEERING

Jayawardhana A.M.S.P.

TEAM Member

IT21838002

it21838002@my.sliit.lk

FACULTY OF COMPUTING | SOFTWARE ENGINEERING

Weerasinghe C.D.

TEAM Member

IT19211688

it19211688@my.sliit.lk

FACULTY OF COMPUTING | SOFTWARE ENGINEERING

Contact Us

Location:

SLIIT Malabe Campus, New Kandy Rd, Malabe, SriLanka

Email:

vocaguideofficial@gmail.com

Call:

+94 70 2648 320