skip to main content
10.1145/3123021.3123060acmconferencesArticle/Chapter ViewAbstractPublication PagesubicompConference Proceedingsconference-collections
research-article

Itchy nose: discreet gesture interaction using EOG sensors in smart eyewear

Published:11 September 2017Publication History

ABSTRACT

We propose a sensing technique for detecting finger movements on the nose, using EOG sensors embedded in the frame of a pair of eyeglasses. Eyeglasses wearers can use their fingers to exert different types of movement on the nose, such as flicking, pushing or rubbing. These subtle gestures can be used to control a wearable computer without calling attention to the user in public. We present two user studies where we test recognition accuracy for these movements.

References

  1. Alonso, W. J., Nascimento, F. C., Shapiro, J., and Schuck-Paim, C. Facing ubiquitous viruses: When hand washing is not enough. Clinical Infectious Diseases 56, 4 (02 2013), 617--617.Google ScholarGoogle ScholarCross RefCross Ref
  2. Ashbrook, D., Tejada, C., Mehta, D., Jiminez, A., Muralitharam, G., Gajendra, S., and Tallents, R. Bitey: An exploration of tooth click gestures for hands-free user interface control. MobileHCI (2016). Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Bedri, A., Byrd, D., Presti, P., Sahni, H., Gue, Z., and Starner, T. Stick it in your ear: Building an in-ear jaw movement sensor. UbiComp/ISWC Adj. (2015). Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Bragi Dash. https://www.bragi.com/thedash/.Google ScholarGoogle Scholar
  5. Bulling, A., Roggen, D., and Tröster, G. Wearable eog goggles: Eye-based interaction in everyday environments. CHI EA (2009). Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Ens, B., Grossman, T., Anderson, F., Matejka, J., and Fitzmaurice, G. Candid interaction: Revealing hidden mobile and wearable computing activities. UIST (2015). Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Gugenheimer, J., Dobbelstein, D., Winkler, C., Haas, G., and Rukzio, E. Facetouch: Enabling touch interaction in display fixed uis for mobile virtual reality. UIST (2016). Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Harrison, C., and Faste, H. Implications of location and touch for on-body projected interfaces. DIS (2014). Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Harrison, C., Tan, D., and Morris, D. Skinput: Appropriating the body as an input surface. CHI (2010). Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Ishimaru, S., Kunze, K., Uema, Y., Kise, K., Inami, M., and Tanaka, K. Smarter eyewear: Using commercial eog glasses for activity recognition. UbiComp Adj. (2014). Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Jones, M., Robinson, S., Pearson, J., Joshi, M., Raju, D., Mbogo, C. C., Wangari, S., Joshi, A., Cutrell, E., and Harper, R. Beyond "yesterday's tomorrow": Future-focused mobile interaction design by and for emergent users. PUC (Feb. 2017). Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Jota, R., and Wigdor, D. Palpebrae superioris: Exploring the design space of eyelid gestures. GI (2015). Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Lissermann, R., Huber, J., Hadjakos, A., and Mühlhäuser, M. Earput: Augmenting behind-the-ear devices for ear-based interaction. CHI EA (2013). Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Manabe, H., and Fukumoto, M. Full-time wearable headphone-type gaze detector. CHI EA (2006). Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Manabe, H., Fukumoto, M., and Yagi, T. Conductive rubber electrodes for earphone-based eye gesture input interface. PUC (Jan. 2015). Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Polacek, O., Grill, T., and Tscheligi, M. Nosetapping: What else can you do with your nose? MUM (2013). Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Rico, J., and Brewster, S. Usable gestures for mobile interfaces: Evaluating social acceptability. CHI (2010). Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Serrano, M., Ens, B. M., and Irani, P. P. Exploring the use of hand-to-face input for interacting with head-worn displays. CHI (2014). Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Apple Watch Users Discover Another Way to Go 'Hands Free'. https://www.wsj.com/articles/nosy-apple-watch-users-discover-another-way-to-go-hands-free-1451077454.Google ScholarGoogle Scholar
  20. Zarek, A., Wigdor, D., and Singh, K. Snout: One-handed use of capacitive touch devices. AVI (2012). Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Zhang, C., Bedri, A., Reyes, G., Bercik, B., Inan, O. T., Starner, T. E., and Abowd, G. D. Tapskin: Recognizing on-skin input for smartwatches. ISS (2016). Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Itchy nose: discreet gesture interaction using EOG sensors in smart eyewear

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader