ABSTRACT
We propose a sensing technique for detecting finger movements on the nose, using EOG sensors embedded in the frame of a pair of eyeglasses. Eyeglasses wearers can use their fingers to exert different types of movement on the nose, such as flicking, pushing or rubbing. These subtle gestures can be used to control a wearable computer without calling attention to the user in public. We present two user studies where we test recognition accuracy for these movements.
- Alonso, W. J., Nascimento, F. C., Shapiro, J., and Schuck-Paim, C. Facing ubiquitous viruses: When hand washing is not enough. Clinical Infectious Diseases 56, 4 (02 2013), 617--617.Google ScholarCross Ref
- Ashbrook, D., Tejada, C., Mehta, D., Jiminez, A., Muralitharam, G., Gajendra, S., and Tallents, R. Bitey: An exploration of tooth click gestures for hands-free user interface control. MobileHCI (2016). Google ScholarDigital Library
- Bedri, A., Byrd, D., Presti, P., Sahni, H., Gue, Z., and Starner, T. Stick it in your ear: Building an in-ear jaw movement sensor. UbiComp/ISWC Adj. (2015). Google ScholarDigital Library
- Bragi Dash. https://www.bragi.com/thedash/.Google Scholar
- Bulling, A., Roggen, D., and Tröster, G. Wearable eog goggles: Eye-based interaction in everyday environments. CHI EA (2009). Google ScholarDigital Library
- Ens, B., Grossman, T., Anderson, F., Matejka, J., and Fitzmaurice, G. Candid interaction: Revealing hidden mobile and wearable computing activities. UIST (2015). Google ScholarDigital Library
- Gugenheimer, J., Dobbelstein, D., Winkler, C., Haas, G., and Rukzio, E. Facetouch: Enabling touch interaction in display fixed uis for mobile virtual reality. UIST (2016). Google ScholarDigital Library
- Harrison, C., and Faste, H. Implications of location and touch for on-body projected interfaces. DIS (2014). Google ScholarDigital Library
- Harrison, C., Tan, D., and Morris, D. Skinput: Appropriating the body as an input surface. CHI (2010). Google ScholarDigital Library
- Ishimaru, S., Kunze, K., Uema, Y., Kise, K., Inami, M., and Tanaka, K. Smarter eyewear: Using commercial eog glasses for activity recognition. UbiComp Adj. (2014). Google ScholarDigital Library
- Jones, M., Robinson, S., Pearson, J., Joshi, M., Raju, D., Mbogo, C. C., Wangari, S., Joshi, A., Cutrell, E., and Harper, R. Beyond "yesterday's tomorrow": Future-focused mobile interaction design by and for emergent users. PUC (Feb. 2017). Google ScholarDigital Library
- Jota, R., and Wigdor, D. Palpebrae superioris: Exploring the design space of eyelid gestures. GI (2015). Google ScholarDigital Library
- Lissermann, R., Huber, J., Hadjakos, A., and Mühlhäuser, M. Earput: Augmenting behind-the-ear devices for ear-based interaction. CHI EA (2013). Google ScholarDigital Library
- Manabe, H., and Fukumoto, M. Full-time wearable headphone-type gaze detector. CHI EA (2006). Google ScholarDigital Library
- Manabe, H., Fukumoto, M., and Yagi, T. Conductive rubber electrodes for earphone-based eye gesture input interface. PUC (Jan. 2015). Google ScholarDigital Library
- Polacek, O., Grill, T., and Tscheligi, M. Nosetapping: What else can you do with your nose? MUM (2013). Google ScholarDigital Library
- Rico, J., and Brewster, S. Usable gestures for mobile interfaces: Evaluating social acceptability. CHI (2010). Google ScholarDigital Library
- Serrano, M., Ens, B. M., and Irani, P. P. Exploring the use of hand-to-face input for interacting with head-worn displays. CHI (2014). Google ScholarDigital Library
- Apple Watch Users Discover Another Way to Go 'Hands Free'. https://www.wsj.com/articles/nosy-apple-watch-users-discover-another-way-to-go-hands-free-1451077454.Google Scholar
- Zarek, A., Wigdor, D., and Singh, K. Snout: One-handed use of capacitive touch devices. AVI (2012). Google ScholarDigital Library
- Zhang, C., Bedri, A., Reyes, G., Bercik, B., Inan, O. T., Starner, T. E., and Abowd, G. D. Tapskin: Recognizing on-skin input for smartwatches. ISS (2016). Google ScholarDigital Library
Index Terms
- Itchy nose: discreet gesture interaction using EOG sensors in smart eyewear
Recommendations
JINSense: Repurposing Electrooculography Sensors on Smart Glass for Midair Gesture and Context Sensing
CHI EA '21: Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing SystemsIn this work, we explore a new sensing technique for smart eyewear equipped with Electrooculography (EOG) sensors. We repurpose the EOG sensors embedded in a JINS MEME smart eyewear, originally designed to detect eye movement, to detect midair hand ...
Automated Data Gathering and Training Tool for Personalized "Itchy Nose"
AH '18: Proceedings of the 9th Augmented Human International ConferenceIn "Itchy Nose" we proposed a sensing technique for detecting finger movements on the nose for supporting subtle and discreet interaction. It uses the electrooculography sensors embedded in the frame of a pair of eyeglasses for data gathering and uses ...
Haptic feedback of gaze gestures with glasses: localization accuracy and effectiveness
UbiComp/ISWC'15 Adjunct: Adjunct Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2015 ACM International Symposium on Wearable ComputersWearable devices including smart eyewear require new interaction methods between the device and the user. In this paper, we describe our work on the combined use of eye tracking for input and haptic (touch) stimulation for output with eyewear. Input ...
Comments