Assistive Technology

A smartphone screen shows a camera-based interface as a person navigates a subway station. The app detects and highlights wayfinding signs in the camera view using OCR and confirms the user is on the correct route. GTFS transit data and real-time sign recognition are combined to provide auditory guidance through complex transit stations.

Commute Booster

Middle mile GTFS OCR Mobile

Commute Booster is a mobile navigation application designed to support people with blindness and low vision during subway travel, particularly within complex transit stations. It combines GTFS-based trip planning with real-time sign recognition using computer vision, enabling users to locate and confirm wayfinding signage inside stations without additional infrastructure.

  • Full trip planning plus real-time station guidance
  • Phone camera sign recognition using OCR
  • GTFS-derived sign list aligned to the intended route
  • Validated in NYC stations with strong accuracy
0.97
Sign accuracy
11 m
Max range
110°
Oblique angle

A person extends their arm and points toward objects in their environment. A wearable sensor tracks the 3D direction of the pointing gesture and identifies the targeted object using computer vision. The system announces the object name aloud, enabling hands-free identification to reduce cognitive load for people with visual impairments.

Point-to-Tell

Gesture control Sensory substitution Object detection 3D pointing

Point-to-Tell 2 is a selective, gesture-controlled assistive system that helps people who are blind or have low vision request only the information they want, reducing extraneous feedback and potential cognitive overload. Using a monocular camera with AI pipelines for depth estimation, hand pose tracking, and object detection/segmentation, the system infers a user's 3D pointing direction and announces the names and distances of pointed-at objects.

  • Pointing-based selection to control what the system describes
  • 3D ray-casting links object identity with spatial position
  • AI stack: depth estimation, hand pose tracking, detection/segmentation
  • Private, hands-free interaction designed for real-world use
High
Hand pose accuracy
Stable
Close-range distance
Bias
Systematic distance error

A person navigates a store aisle while holding a smartphone running the ShopScout app. The system detects product labels and reads barcodes in real time using object detection. Spatial audio cues guide the shopper toward target items and vision-language models provide product descriptions, enabling fully independent shopping.

ShopScout

Assistive technology Object detection Vision-language Spatial audio

ShopScout is a multimodal assistive system that enables people with blindness and low vision to shop independently. By combining object detection, vision-language models, and spatial audio feedback, the system guides users from shopping list creation to successful item retrieval without requiring assistance.

  • End-to-end workflow from list creation to item retrieval
  • Real-time object detection for product identification
  • Vision-language models for contextual understanding
  • Spatial audio feedback for intuitive navigation
Multi
Modal integration
Real-time
Feedback
Independent
Shopping

A person walks through an environment while wearing a haptic belt around their waist. Camera detects nearby obstacles or walkable space and translate proximity into directional vibrations on the belt. The system provides continuous sensory substitution feedback, guiding the user along open paths without relying on vision.

Virtual Whiskers wearable haptic belt

Virtual Whiskers

Haptics Sensory substitution Obstacle negotiation Wearable

Virtual Whiskers is a haptics-based electronic travel aid designed to support obstacle negotiation for people with blindness and low vision. Using modular vibration units and higher-order sensory substitution, the system conveys navigational information through adaptive vibrotactile feedback.

  • Modular vibration units operate independently for spatial feedback
  • Open path mode guides users toward the most traversable direction
  • Depth mode maps proximity to vibration intensity
  • User testing with participants with blindness or low vision
10
Participants
Reduced
Hesitation
Fewer
Cane contacts

A person walks through inside a building using a phone camera that performs visual place recognition in real time. Without GPS or fixed beacons, the system estimates position and heading using stored visual maps. On-screen overlays display estimated location and navigation instructions derived from the camera feed.

UNAV

Vision-based localization VPR PnP Topometric map

UNAV is a vision-based localization and navigation pipeline designed for end users with blindness and low vision. Given a query image captured in a mobile application, the system performs visual place recognition to retrieve similar reference images and estimates the user's location via a weighted-average approach. It also estimates user heading using a perspective-n-point method based on 2D–3D correspondences, then computes a shortest path on a navigable map to support end-to-end wayfinding.

  • Visual place recognition retrieves similar images from a reference database
  • Location estimated using a weighted-average method on matched image geolocations
  • Heading estimated via PnP using 2D–3D point correspondences
  • Topometric map built by projecting a 3D sparse map onto a 2D floor plan
  • Shortest path computed with Dijkstra's algorithm on a navigable map
< 1 m
Avg localization error
No
Camera intrinsics needed
Tested
Hospital environment

A wearable camera scans the sidewalk ahead as a person approaches a curb. The real-time vision system segments the pavement and identifies the curb edge before it is reached. An auditory alert is delivered in advance, giving the user time to orient and step safely, improving outdoor mobility for people with visual impairments.

Curb Detector

Pedestrian safety Object detection Sensory substitution Wearable

Curb Detector is a vision-based wearable assistive system designed to help people who are blind or have low vision safely detect and orient to curbs in urban environments. Using an RGB camera and an embedded processing platform, the system segments curbs in real time and provides early warning and orientation information before the user reaches an elevation change.

  • Real-time curb segmentation using a YOLOv8 model trained on a custom dataset
  • Adaptive auditory feedback including beeps, abstract sonification, and speech
  • Advanced warning with a larger safety window than a white cane
  • Provides curb distance and orientation information
YOLOv8
Segmentation model
Earlier
Curb warning
Comparable
Orientation vs cane

A smartphone camera scans a city sidewalk where construction work is underway. Bounding boxes appear around detected hazards including barriers, scaffolding, and warning signs, identified using YOLO object detection models. OCR reads visible signage text and the system announces hazard information audibly to help users navigate safely around construction zones.

Construction Site Detection

Urban navigation Computer vision Hazard detection Assistive technology

Construction Site Detection is a real-time, computer-vision–based assistive system designed to help people who are blind or have low vision identify temporary construction hazards in urban environments. The system provides advance warning of construction zones that disrupt familiar routes and introduce safety risks such as uneven surfaces, barriers, and altered walkways.

  • Open-vocabulary detection of diverse construction elements
  • YOLO-based model specialized for scaffolding and poles
  • OCR interprets construction signage
  • Validated in static sites and dynamic walking routes
88.6%
Static accuracy
92.0%
Dynamic accuracy
2–10 m
Detection range

Sensorimotor Integration

KineReach visuomotor coordination experimental setup with integrated eye tracking

KineReach With Integrated Eye-Tracking

Visuomotor control Sensorimotor integration Kinematics Eye tracking

KineReach is a high-precision, time-synchronized measurement and analysis pipeline for quantifying visuomotor coordination during goal-directed behavior. Using controlled paradigms that dissociate gaze from reach planning, the platform models eye–hand timing, variability, and online correction to reveal mechanistic signatures of sensorimotor function, impairment, and recovery.

  • High-precision synchronization of gaze, limb kinematics, and task events
  • Paradigms separating planning from execution
  • Trial-level eye–hand latency and coordination variability
  • Mechanistic biomarkers responsive to rehabilitation
ms
Temporal precision
Trial
Granularity
Sensitive
To impairment
Neural-muscular connectivity analysis for motor recovery profiling

Neural-Muscular Connectivity in Motor Recovery

Biosignal Fusion Network Analysis Neurorehabilitation

This project investigates neural-muscular coupling as a biomarker for motor recovery following neurological events. By fusing brain and muscle activity signals, we construct connectivity networks and extract graph-theoretic features to capture recovery trajectories beyond conventional clinical scales.

  • Fused brain-muscle connectivity network construction
  • Graph-theoretic biomarkers for recovery profiling
  • Correlation with functional and kinematic outcomes
  • Cross-sectional and longitudinal study design
Fused
Biosignals
Network
Biomarkers
Recovery
Profiling

A participant sits at a table performing a standardized hand function task while wearing eye-tracking glasses. Cameras and motion capture sensors record fine finger movements and gaze patterns simultaneously. The multimodal system extracts objective biomarkers of movement strategy and quality to support sensitive rehabilitation assessment.

Upper extremity motor assessment

Rehabilitation Motion capture Eye tracking Functional biomarkers

This project enhances standardized clinical hand-function assessments by integrating multimodal sensing pipelines. We derive objective biomarkers related to movement strategy, quality, and error patterns—going well beyond traditional completion-time measures to support sensitive phenotyping and rehabilitation-focused evaluation.

  • Synchronized gaze and motion capture during functional tasks
  • Biomarkers beyond completion time
  • Subtask-level event timing and kinematics
  • Quantitative outcomes for longitudinal tracking
Multimodal
trial metrics
Strategy
Profiling
Clinical
Relevance

A person performs the Nine-Hole Peg Test under overhead camera observation. The computer vision system tracks hand and finger movements frame-by-frame using object detection and pose estimation. Objective performance metrics beyond simple completion time are extracted to quantify fine motor impairment in individuals with Multiple Sclerosis.

Multiple Sclerosis Assessment

Computer vision Multiple sclerosis Hand tracking Eye tracking

This project augments standard clinical dexterity assessments with advanced computer vision pipelines to objectively quantify fine motor performance in individuals with neurological conditions. By fusing object detection with hand tracking, the system produces granular performance metrics beyond simple completion time.

  • Augments standard assessments using vision-based pipelines
  • Fusion of object detection and eye-hand tracking
  • Quantifiable, repeatable datasets
  • Robust detection and error handling
Vision
Driven Quantification
Beyond
Completion time
Scalable
Pipeline

A child's hands are captured by multiple cameras as they perform a fine motor task. Neural network-based pose estimation reconstructs 3D finger movements without any wearable markers or sensors. The system provides millimeter-scale spatial data to support precise diagnostics and rehabilitation planning for children with neurological injuries.

Pediatric Rehabilitation Hand Tracking

Computer vision Pediatric rehabilitation Hand tracking Eye tracking

This project develops a markerless, multi-camera pipeline for reconstructing 3D hand kinematics during dexterity tasks in children. Using neural network-based pose estimation and multi-view reconstruction, it captures fine finger movements without wearable sensors, overcoming key limitations of traditional motion capture. Paired with modern eye-tracking, this framework supports more precise diagnostics and targeted rehabilitation for children with neurologic injury.

  • Markerless multi-camera 3D hand reconstruction
  • Task-specific neural network training for pediatric hands
  • Millimeter-scale spatial reconstruction without sensors
  • Compatibility with integrated eye-tracking systems
Markerless
3D Reconstruction
Multi-view
Camera Pipeline
Scalable
for clinical workflow