Man Versus Machine Reaches Symbiotic State; Eye Tracking and Gesture Applications to Revolutionize Patient Care
Oyster Bay, New York - 11 Jul 2016
Gesture, eye tracking, and proximity sensor technologies will mark the next stage of innovation in machine design, finds ABI Research. A broader and more competitive ecosystem spurred by smartphone and tablet sensor integration—forecast to hit close to $5 billion in 2016—will create massive opportunities in automotive, consumer electronics, and healthcare. Healthcare, in particular, shows the largest, untapped opportunity for eye tracking and gesture applications in patient care.
“The same way that touchscreens eclipsed the PC mouse, gesture and eye tracking sensors will transform the way people interact with machines, systems, and their environment,” says Jeff Orr, Research Director for ABI Research. “Healthcare professionals are relying on these sensors to move away from subjective patient observations and toward more quantifiable and measurable prognoses, revolutionizing patient care.”
Eye tracking sensors can help detect concussions and head trauma, identify autism in children before they are speaking, and enable vision therapy programs for early childhood learning challenges to retrain the learned aspects of vision. Similarly, gesture sensors are translating sign language into speech, providing doctors a means to manipulate imaging hands-free during surgical procedures, and providing a natural means to navigate through virtual experiences.
Both established and startup companies are involved in the human-machine interface revolution. Sensor innovation is stemming from Hillcrest Labs, NXP, and Synaptics, among others. Atheer, Bluemint Labs, eyeSight, Google, Intel, Leap Motion, Microsoft, Nod Labs, RightEye, and Tobii Group also all recently announced creative gesture, proximity, and eye tracking solutions.
“Healthcare is only one industry poised to benefit from reinventing the user interface,” adds Orr. “The larger competitive ecosystem for perceptual sensors is forging opportunities in consumer appliances, autonomous driving, musical instruments, gaming, retail, and even hazardous locations.”
These findings are from ABI Research’s Eye Tracking, Gestures and Proximity Sensor Applications and Human-Machine Interfaces webinar. This report is part of the company’s Wearables & Devices sector, which includes research, data, and analyst insights.
Contact ABI Research
- Americas +1.516.624.2542
- Europe +44.(0).203.326.0142
- Asia +65.6592.0290
Other Recent News
- 21 Mar 2019
- Stratasys Takes Top Spot in ABI Research’s Additive Manufacturing Platform Competitive Assessment
- 20 Mar 2019
- ABI Research Ranks Brain Corp, Seegrid, Balyo, and ASI as Top Technology Providers in the Mobile Robot Space
- 19 Mar 2019
- China’s Commercial Telematics Market Evolving, Nearing $5 Billion in System Revenues by 2024
- 14 Mar 2019
- Fingerprint Sensor Shipments to reach 1.26 Billion in 2019 Despite Growth from Face Recognition