MAHITA is an intelligent visual assistant designed to improve mobility and safety for blind and visually impaired people. It combines a smartphone camera, on-device AI, and an ultrasonic sensor to detect obstacles and describe the surrounding environment in real time.

The system is fully accessible through voice commands (STT), audio responses (TTS), and haptic alerts (vibration/buzzer), helping users avoid collisions, understand nearby objects, and move more confidently during daily travel.

Client:

Student Team Project

Industry:

MISA Antananarivo

Duration:

3 Months

Flutter WebSocket Google MLKit TensorFlow Lite BLIP ESP32 C++ Electronics

Problem-Solving Overview

From identifying issues to delivering solutions with impact.

Problem

Problem

  • Lack of autonomy for visually impaired people
  • Limited perception of the environment
  • Uncertain and stressful mobility
Causes

Causes

  • Lack of suitable tools
  • Outdated tools (e.g., white cane)
  • Unpredictable urban environments
  • Unintuitive interfaces
Consequences

Consequences

  • Unexpected collisions
  • Stress
  • Increased dependence
  • Social isolation
Solution

Solution

  • Wearable device
  • Embedded AI
  • Ultrasonic sensors
  • Voice guidance
  • Real-time communication

Key Features

Discover the core functionalities and how they improve user experience.

Image description by AI
Feature 01

Image Description

Mahita analyzes the environment captured by the camera and describes the objects and scenes present.

Google MLKit (Image Labeling) BLIP
Facial recognition
Feature 02

Facial Recognition

Identifies known faces stored in the app's database so the user can recognize people nearby.
Users can add new faces by capturing images and associating them with names.

MLKit Face Detection face_recognition.tflite SQLite
Text translation
Feature 03

Text Translation

Recognizes text in images or from voice and translates it to or from a foreign language, then reads it aloud.

Google MLKit OCR Language Identification MLKit Translation Text-to-Speech
Obstacle detection
Feature 04

Obstacle Detection

Measures the distance to nearby objects and alerts the user to avoid collisions via audio feedback and buzzer alerts.

HC-SR04 ESP32 Buzzer
Physical button interaction
Feature 05

Physical Button Interaction

Allows the user to navigate the app or trigger actions without using the screen.

Physical Buttons ESP32 WebSocket
Voice feedback TTS
Feature 06

Voice Feedback (Text-to-Speech)

All information (image description, facial recognition, translation, alerts) is delivered vocally to the user.

Flutter TTS
Send SMS
Feature 07

Send SMS

Send the spoken message via SMS to a selected contact.

sms_sender

My Role in the Project

What i really did to make this project a success.

Mobile App development

  • Designed and implemented the mobile app using Flutter.
  • Integrated all Google MLKit features.
  • Integrated embedded AI using TensorFlow Lite.
  • Implemented WebSocket communication for real-time data exchange with ESP32.
  • Translation features for multiple languages.
  • Sent SMS messages to selected contacts.
  • Implemented voice command features using Speech-to-Text (STT).
  • Developed Text-to-Speech (TTS) for audio feedback.
image loading...

I enjoy discussing new projects challenges. Tell me about your goals, your needs, and your timeline and let’s build something great together.