MediVox, Voice-recognition User Interface
This Voice-recognition User Interface product was designed and developed over a 48-hour period at the 2016 Launch Hackathon in San Francisco. By the end we had a working MVP (Minimum Viable Product) to demo that captured the biometrics, symptoms, and notes. View the video below to a sample of the interaction design.
Healthcare providers are missing out on key facial, body language, and conversational interactions with their patients while they are heads-down entering data into Electronic Health Records (EHR, EMR).
The Solution: a voice user interface with machine-learning
My team of three at Launch Hackathon was comprised of a front-end engineer, a back-end database expert, and product designer. We used voice user interface and machine-learning to let doctors and patients talk conversationally, capture that conversation into the appropriate areas of biometrics, symptoms, prescriptions, labs, next appointment, and notes, display the information in real-time, allowed for edits via voice (no keyboard needed), and then saved it in the patient’s EHR (Electronic Health Record). Future iterations were planned to let prescriptions be created and sent to their pharmacy, labs ordered, and followup appointments made. As this interface was largely voice, we kept the visual interface simple and intuitive.
The integration plan for this platform was to integrate into the patient’s permanent EHR – simplifying a user interface that would keep health care providers and their patients happy.