[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3342558.3345400acmconferencesArticle/Chapter ViewAbstractPublication PagesdocengConference Proceedingsconference-collections
research-article

Modeling Multimodal-Multiuser Interactions in Declarative Multimedia Languages

Published: 23 September 2019 Publication History

Abstract

Recent advances in hardware and software technologies have given rise to a new class of human-computer interfaces that both explores multiple modalities and allows for multiple collaborating users. When compared to the development of traditional single-user WIMP (windows, icons, menus, pointer)-based applications, however, applications supporting the seamless integration of multimodal-multiuser interactions bring new specification and runtime requirements. With the aim of assisting the specification of multimedia applications that integrate multimodal-multiuser interactions, this paper: (1) proposes the MMAM (Multimodal-Multiuser Authoring Model); (2) presents three different instantiations of it (in NCL, HTML, and a block-based syntax); and (3) evaluates the proposed model through a task-based user study. MMAM enables programmers to design and ponder different solutions for applications with multimodal-multiuser requirements. The proposed instantiations served as proofs of concept about the feasibility of our model implementation and provided the basis for practical experimentation, while the performed user study focused on capturing evidence of both the user understanding and the user acceptance of the proposed model. We asked developers to perform tasks using MMAM and then answer a TAM (Technology Acceptance Model)-based questionnaire focused on both the model and its instances. As results, the study indicates that the participants easily understood the model (most of them performed the required tasks with minor or no errors) and found it both useful and easy to use. 94.47% of the participants gave positive answers to the block-based representation TAM questions, whereas 75.17% gave positive answers to the instances-related questions.

References

[1]
Richard A. Bolt. 1980. Put-That-There: Voice and Gesture at the Graphics Interface. In Proceedings of the 7th Annual Conference on Computer Graphics and Interactive Techniques. ACM, New York, NY, USA, 262--270. https://doi.org/10.1145/800250.807503
[2]
Lucas Carvalho and Hendrik Macedo. 2010. Estendendo a NCL para Promover Interatividade Vocal em Aplicações Ginga na TVDi Brasileira. In WebMedia '10: Proceedings of the 16th Brazilian Symposium on Multimedia and the Web. Proceedings of the XIV Brazilian Symposium on Multimedia and the Web.
[3]
Deborah A Dahl. 2017. Multimodal Interaction with W3C Standards. Springer.
[4]
Fred D. Davis. 1989. Perceived Usefulness, Perceived Ease of Use, and User Acceptance of Information Technology. MIS Quarterly 13, 3 (1989), 319--340. https://doi.org/10.2307/249008
[5]
Prasun Dewan and Rajiv Choudhary. 1991. Primitives for Programming Multiuser Interfaces. In Proceedings of the 4th Annual ACM Symposium on User Interface Software and Technology (UIST '91). ACM, New York, NY, USA, 69--78. https://doi.org/10.1145/120782.120790
[6]
Bruno Dumas, Denis Lalanne, and Rolf Ingold. 2010. Description languages for multimodal interaction: a set of guidelines and its illustration with SMUIML. Journal on Multimodal User Interfaces 3, 3 (April 2010), 237--247. https://doi.org/10.1007/s12193-010-0043-3
[7]
Alan Esenther, Cliff Forlines, Kathy Ryall, and Sam Shipman. 2002. Diamondtouch sdk: Support for multi-user, multi-touch applications. Mitsubishi Electric Research Laboratories, Inc., five pages (2002).
[8]
Luiz Fernando and Gomes Soares. 2009. O uso da linguagem declarativa do ginga-ncl na construção de conteúdos audiovisuais interativos: a experiência do "roteiros do dia". In Proceedings at 1° Simpósio Internacional de Televisão Digital.
[9]
Gregg Stanley Foster. 1986. Collaborative Systems and Multi-user Interfaces. PhD Thesis. University of California, Berkeley.
[10]
David Gefen and Mark Keil. 1998. The Impact of Developer Responsiveness on Perceptions of Usefulness and Ease of Use: An Extension of the Technology Acceptance Model. SIGMIS Database 29, 2 (April 1998), 35--49. https://doi.org/10.1145/298752.298757
[11]
Gheorghita Ghinea, Christian Timmerer, Weisi Lin, and Stephen R. Gulliver. 2014. Mulsemedia: State of the Art, Perspectives, and Challenges. ACM Trans. Multimedia Comput. Commun. Appl. 11, 1s, Article 17 (Oct. 2014), 23 pages. https://doi.org/10.1145/2617994
[12]
Alan L. V. Guedes, Marcio Cunha, Hugo Fuks, Sérgio Colcher, and Simone Diniz Junqueira Barbosa. 2016. Using NCL to Synchronize Media Objects, Sensors and Actuators. http://www.lbd.dcc.ufmg.br/colecoes/wsotwebmedia/2016/003.pdf
[13]
Alan L. V. Guedes, Roberto G. de Albuquerque Azevedo, Sérgio Colcher, and Simone D.J. Barbosa. 2016. Extending NCL to Support Multiuser and Multimodal Interactions. In Proceedings of the 22Nd Brazilian Symposium on Multimedia and the Web (Webmedia '16). ACM, 39--46. https://doi.org/10.1145/2976796.2976869
[14]
Josefina Guerrero Garcia, Jean Vanderdonckt, and others. 2010. Designing workflow user interfaces with UsiXML. In 1st Int. Workshop on User Interface eXtensible Markup Language UsiXML'2010.
[15]
Cornelia Haber. 2001. Modeling Multiuser Interactions. In Proceedings at the First European Computer Supported Collaborative Learning Conference, Maastricht, Germany. 22--24.
[16]
ITU. 2009. H.761: Nested Context Language (NCL) and Ginga-NCL for IPTV Services. Technical Report. ITU, Geneva, Switzerland.
[17]
Alejandro Jaimes and Nicu Sebe. 2007. Multimodal human-computer interaction: A survey. Computer Vision and Image Understanding 108, 1 (Oct. 2007), 116--134. https://doi.org/10.1016/j.cviu.2006.10.019
[18]
Marina Josué, Raphael Abreu, Fábio Barreto, Douglas Mattos, Glauco Amorim, Joel dos Santos, and Débora Muchaluat-Saade. 2018. Modeling Sensory Effects As First-class Entities in Multimedia Applications. In Proceedings of the 9th ACM Multimedia Systems Conference (MMSys '18). ACM, New York, NY, USA, 225--236. https://doi.org/10.1145/3204949.3204967
[19]
Yann Laurillau. 2013. IOWAState: implementation models and design patterns for identity-aware user interfaces based on state machines. In Proceedings of the 5th ACM SIGCHI symposium on Engineering interactive computing systems. ACM, 59--68.
[20]
Dirk Schnelle-Walka, Stefan Radomski, and Max Mühlhäuser. 2013. JVoiceXML as a modality component in the W3C multimodal architecture. Journal on Multimodal User Interfaces 7, 3 (Nov. 2013), 183--194. https://doi.org/10.1007/s12193-013-0119-y
[21]
R. Benjamin Shapiro and Matthew Ahrens. 2016. Beyond Blocks: Syntax and Semantics. Commun. ACM 59, 5 (April 2016), 39--41. https://doi.org/10.1145/2903751
[22]
Carlos de Salles Soares Neto, Luiz Fernando Gomes Soares, and Clarisse Sieckenius de Souza. 2010. The Nested Context Language reuse features. Journal of the Brazilian Computer Society 16, 4 (01 Nov 2010), 229--245. https://doi.org/10.1007/s13173-010-0017-z
[23]
Edward Tse, Saul Greenberg, Chia Shen, and Clifton Forlines. [n.d.]. Multimodal Multiplayer Tabletop Gaming. 5, 2 ([n. d.]). https://doi.org/10.1145/1279540.1279552
[24]
Edward Tse, Saul Greenberg, Chia Shen, Clifton Forlines, and Ryo Kodama. 2008. Exploring True Multi-user Multimodal Interaction over a Digital Table. In Proceedings of the 7th ACM Conference on Designing Interactive Systems (DIS '08). ACM, New York, NY, USA, 109--118. https://doi.org/10.1145/1394445.1394457
[25]
Matthew Turk. 2014. Multimodal interaction: A review. Pattern Recognition Letters 36 (2014), 189--195. https://doi.org/10.1016/j.patrec.2013.07.003
[26]
W3C. 2001. XHTML+Voice Profile 1.0. http://www.w3.org/TR/xhtml+voice
[27]
W3C. 2012. Multimodal Architecture and Interfaces. http://www.w3.org/TR/mmi-arch

Cited By

View all
  • (2020)Interactive 360-degree Videos in Ginga-NCL Using Head-Mounted-Displays as Second Screen DevicesProceedings of the Brazilian Symposium on Multimedia and the Web10.1145/3428658.3430972(289-296)Online publication date: 30-Nov-2020

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
DocEng '19: Proceedings of the ACM Symposium on Document Engineering 2019
September 2019
254 pages
ISBN:9781450368872
DOI:10.1145/3342558
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

In-Cooperation

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 23 September 2019

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. HTML
  2. MUI
  3. Multimedia Languages
  4. Multimodal User Interactions
  5. Multiuser User Interactions
  6. NCL

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

DocEng '19
Sponsor:
DocEng '19: ACM Symposium on Document Engineering 2019
September 23 - 26, 2019
Berlin, Germany

Acceptance Rates

DocEng '19 Paper Acceptance Rate 30 of 77 submissions, 39%;
Overall Acceptance Rate 194 of 564 submissions, 34%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)5
  • Downloads (Last 6 weeks)0
Reflects downloads up to 12 Dec 2024

Other Metrics

Citations

Cited By

View all
  • (2020)Interactive 360-degree Videos in Ginga-NCL Using Head-Mounted-Displays as Second Screen DevicesProceedings of the Brazilian Symposium on Multimedia and the Web10.1145/3428658.3430972(289-296)Online publication date: 30-Nov-2020

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media