Change your cover photo
Upload
Change your cover photo
I'm a third-year Electronics and Communication Engineering student at RVCE. Skilled in C, Python, embedded C, MATLAB, LTspice, and machine learning.
This user account status is Approved

This user has not added any information to their profile yet.

Vision-Enhanced Mobility: Real-time Environmental Recognition and Location Tracking
Maixduino Kit for AI+IoT

The proposed system integrates advanced technologies to enhance mobility and safety for visually impaired individuals. It utilizes a GPS module to fetch real-time location data of the blind person, sending this information to a server or cloud for access by close ones. The environment is recognized by capturing images with a camera module, and a neural network processes these images to identify objects and surroundings, displaying this information on a display module for easy understanding by the close ones. Crucial details such as the blind person's location, name, and blood group are shown on the display, sourced from a pre-defined database or cloud storage. The system also includes a microphone for receiving voice commands or feedback, and it employs text-to-speech functionality to relay recognized environment details to the blind person. Seamless integration between all modules is ensured, with communication protocols like Wi-Fi and Bluetooth connecting the ESP32 board to the cloud/server for real-time data transmission.

Components Required:-
1. Display Module
2. Camera Module
3. Microphone
4. Wi-Fi
5. Speaker (for audio feedback)
6. Ultrasonic sensor
7. Buzzer
8. Sipeed Maixduino MCU

Methodology:-

The blind person guidance system integrates several key modules on an ESP32 development board with an AI accelerator (Sipeed Maixduino MCU). Using GPS and Wi-Fi connectivity, the system fetches and communicates the blind person's real-time location to caregivers. A camera module equipped with neural networks detects and analyzes the environment, identifying objects and obstacles. Information such as the blind person's location, name, and medical details is displayed on a dedicated module for caregivers. Simultaneously, a microphone captures surrounding sounds, while an AI algorithm processes this data to provide spoken feedback through a speaker, offering navigation directions and alerting to potential hazards. Additional features include ultrasonic sensors for obstacle detection and a buzzer for immediate alerts, ensuring comprehensive assistance and safety during navigation.

Abdulmannan Lalshawala
Gaurav Makhija