Visual impairment represents a critical challenge for our society with 285 million affect worldwide; alarmingly, the prevalence is expected to triple by 2050. Supporting mobility is a chief priority for assistive technologies. In recent years, the integration of computer vision and haptic technologies has led to a number of wearable electronic travel aids (ETAs). Previously, we have proposed an ETA comprised of a computer vision system and a wearable haptic device in the form of a belt. The belt encompasses a two-by-five array of piezoelectric-based macro-fiber composite (MFC) actuators, which can generate vibrations on the abdomen when an oscillating voltage is applied across their electrodes. The computer vision system identifies position and distance of surrounding obstacles egocentrically and drives the actuators relative to the salience of the potential hazard(s). Despite promising pilots, the design, control, and optimization of the ETA requires substantial, potentially high-risk, and tedious training and testing to accommodate patient-specific behavioral idiosyncrasies and a variety of visual impairments. To address these issues, we employ a virtual reality (VR) platform that offers simulations of visual impairment by disease type and severity with front-end control. We review our early work on the first three visual impairments piloted in the platform, each with three levels of severity: mild, moderate and severe. The VR environment is interfaced with the ETA, which provides feedback to the user based on the position of virtual obstacles. These simulations allow safe, controlled, repeatable experiments with ETAs that can be performed with varying degrees of visual perception. Our framework can become a paradigm for the development and testing of ETAs, with other potential applications in disability awareness, education, and training.