ThirdEye

Contribution Project

AI-Powered Spatial Navigator for the Blind people.

Description

ThirdEye AI is an advanced assistive technology designed to bridge the gap between visually impaired individuals and their environment. While traditional aids like white canes detect ground-level physical obstacles, they fail to provide "contextual intelligence." ThirdEye AI uses Computer Vision and Deep Learning to act as a digital pair of eyes, narrating the world in real-time. It transforms raw visual data into descriptive audio, enabling users to identify people, read text, recognize currency, and perceive emotions hands-free.

The Problem

​Global statistics indicate that over 250 million people suffer from moderate to severe vision impairment. Current navigation aids are primitive:

  • Contextual Blindness: A cane can detect a wall, but it cannot tell the user that the "wall" is actually a person they know or a closed door with a "Pull" sign.

  • Financial & Social Isolation: Inability to identify currency notes or read facial expressions leads to a lack of independence and social anxiety.

  • Information Gap: Accessing printed information (menus, medicine labels, signs) remains a significant barrier to daily autonomThe Solution (The Innovation)

ThirdEye AI is a wearable/smartphone-integrated platform that leverages the YOLO (You Only Look Once) architecture for real-time object detection and OCR (Optical Character Recognition) for text extraction.

Key Technical Pillars:

  • Spatial Intelligence: Unlike standard apps, ThirdEye uses a Vertical Grid Mapping system to inform the user of an object's precise location (e.g., "Person on your left" vs. "Door straight ahead").

  • Cognitive Analysis: Beyond object detection, it performs Sentiment Analysis to detect human emotions, helping users navigate social cues.

  • Currency Verification: A dedicated module for Indian Currency (INR) that uses color-signature and watermark detection to prevent financial fraud.

  • Edge Computing Optimized: The system is designed for low-latency performance, ensuring audio feedback is delivered in near real-time (<200ms).

Unique Features (The "Wow" Factor)

  1. Hands-Free Wearable Design: Can be integrated with a smartphone chest-mount or smart-glasses, allowing the user to keep their hands free for a cane or balance.

  2. Scene Description Engine: Instead of just naming objects, it describes scenes (e.g., "A busy street with a car approaching from the right").

  3. Haptic-Audio Feedback: Uses directional audio (Spatial Audio) so the user intuitively knows which direction to turn.

  4. Offline Capability: Core navigation and object detection features are optimized to work without an active internet connection for maximum reliability.

Issues & PRs Board
No issues or pull requests added.