[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/2380296.2380338acmconferencesArticle/Chapter ViewAbstractPublication PagesuistConference Proceedingsconference-collections
poster

MISO: a context-sensitive multimodal interface for smart objects based on hand gestures and finger snaps

Published: 07 October 2012 Publication History

Abstract

We present an unobtrusive multimodal interface for smart objects (MISO) in an everyday indoor environment. MISO uses pointing for object selection and context-sensitive arm gestures for object control. Finger snaps are used to confirm object selections and to aid with gesture segmentation. Audio feedback is provided during the interaction. The use of a Kinect depth camera allows for a compact system and robustness in varying environments and lighting conditions at low cost.

Supplementary Material

JPG File (0226-file3.jpg)
suppl.mov (0226-file3.mp4)
Supplemental video

References

[1]
Caon, M., Yue, Y., Tscherrig, J., Mugellini, E., and Khaled, O. A. Context-Aware 3D Gesture Interaction Based on Multiple Kinects. In AMBIENT 2011 (2011).
[2]
Do, J.-H., Jung, S. H., Jang, H., Yang, S.-E., Jung, J.-W., and Bien, Z. Gesture-Based Interface for Home Appliance Control in Smart Home. In Smart Homes and Beyond - ICOST2006 4th Int. Conference On Smart homes and health Telematics, C. Nugent and J. C. Augusto, Eds. IOS Press, Amsterdam, 2006, 23--30.
[3]
Fleer, D. Steuerung von Umgebungsintelligenz mit multimodalen Gesten. Master's thesis, Universität Bielefeld, 2011.
[4]
Kriesten, B., Tünnermann, R., Mertes, C., and Hermann, T. Controlling ambient information flow between smart objects with a mobile mixed-reality interface. In Proc. of the 12th Int. Conference on Human computer interaction with mobile devices and services, MobileHCI '10, ACM (New York, NY, USA, 2010), 405--406.
[5]
Lin, S.-Y., Lai, Y.-C., Chan, L.-W., and Hung, Y.-P. Real-Time 3D Model-Based Gesture Tracking for Multimedia Control. In 20th Int. Conference on Pattern Recognition (2010).
[6]
Vesa, S., and Lokki, T. An eyes-free user interface controlled by finger snaps. Proc. of the 8th Int. Conference on Digital Audio Effects (DAFx.05) (2005), 262--265.

Cited By

View all
  • (2022)Investigating Clutching Interactions for Touchless Medical Imaging SystemsProceedings of the 2022 CHI Conference on Human Factors in Computing Systems10.1145/3491102.3517512(1-14)Online publication date: 29-Apr-2022
  • (2020)Non-Verbal Auditory Input for Controlling Binary, Discrete, and Continuous Input in Automotive User InterfacesProceedings of the 2020 CHI Conference on Human Factors in Computing Systems10.1145/3313831.3376816(1-13)Online publication date: 21-Apr-2020
  • (2018)UbiCompassAdvances in Human-Computer Interaction10.1155/2018/57813632018Online publication date: 12-Apr-2018
  • Show More Cited By

Index Terms

  1. MISO: a context-sensitive multimodal interface for smart objects based on hand gestures and finger snaps

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    UIST Adjunct Proceedings '12: Adjunct proceedings of the 25th annual ACM symposium on User interface software and technology
    October 2012
    114 pages
    ISBN:9781450315821
    DOI:10.1145/2380296

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 07 October 2012

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. ambient intelligence
    2. finger snaps
    3. gestures
    4. home automation
    5. multimodal interfaces

    Qualifiers

    • Poster

    Conference

    UIST '12

    Acceptance Rates

    Overall Acceptance Rate 355 of 1,733 submissions, 20%

    Upcoming Conference

    UIST '25
    The 38th Annual ACM Symposium on User Interface Software and Technology
    September 28 - October 1, 2025
    Busan , Republic of Korea

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)11
    • Downloads (Last 6 weeks)4
    Reflects downloads up to 01 Jan 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2022)Investigating Clutching Interactions for Touchless Medical Imaging SystemsProceedings of the 2022 CHI Conference on Human Factors in Computing Systems10.1145/3491102.3517512(1-14)Online publication date: 29-Apr-2022
    • (2020)Non-Verbal Auditory Input for Controlling Binary, Discrete, and Continuous Input in Automotive User InterfacesProceedings of the 2020 CHI Conference on Human Factors in Computing Systems10.1145/3313831.3376816(1-13)Online publication date: 21-Apr-2020
    • (2018)UbiCompassAdvances in Human-Computer Interaction10.1155/2018/57813632018Online publication date: 12-Apr-2018
    • (2018)Tap-to-PairProceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies10.1145/32870792:4(1-21)Online publication date: 27-Dec-2018
    • (2016)AmbiGazeProceedings of the 2016 ACM Conference on Designing Interactive Systems10.1145/2901790.2901867(812-817)Online publication date: 4-Jun-2016
    • (2016)Snap-To-ItProceedings of the 2016 CHI Conference on Human Factors in Computing Systems10.1145/2858036.2858177(5909-5920)Online publication date: 7-May-2016
    • (2016)Personal Smart Space: IoT Based User Recognition and Device Control2016 IEEE Tenth International Conference on Semantic Computing (ICSC)10.1109/ICSC.2016.45(181-182)Online publication date: Feb-2016
    • (2015)Where, what, why and how - 3W1HProceedings of the 14th Brazilian Symposium on Human Factors in Computing Systems10.1145/3148456.3148472(1-10)Online publication date: 3-Nov-2015
    • (2013)Point & control -- interaction in smart environmentsProceedings of the 2013 ACM conference on Pervasive and ubiquitous computing adjunct publication10.1145/2494091.2494184(303-306)Online publication date: 8-Sep-2013

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media