[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3308557.3308730acmconferencesArticle/Chapter ViewAbstractPublication PagesiuiConference Proceedingsconference-collections
short-paper

A modular framework for collaborative multimodal annotation and visualization

Published: 16 March 2019 Publication History

Abstract

Artificial Intelligence (AI) research, including machine learning, computer vision, and natural language, requires large amounts of annotated datasets. The current research and development (R&D) pipeline involves each group collecting their own datasets using an annotation tool tailored specifically to their needs, followed by a series of engineering efforts in loading other external datasets and developing their own interfaces, often mimicking some components of existing annotation tools. In departure from the current paradigm, my research focuses on reducing inefficiencies by developing a unified web-based, fully configurable framework that enables researchers to set up an end-to-end R&D experience from dataset annotations to deployment with an application-specific AI backend. Extensible and customizable as required by individual projects, the framework has been successfully featured in a number of research efforts, including conversational AI, explainable AI, and commonsense grounding of language and vision. This submission outlines the various milestones-to-date and planned future work.

References

[1]
Eckart et al. 2016. A Web-based Tool for the Integrated Annotation of Semantic and Syntactic Structures. LT4DH COLING-Workshop (2016).
[2]
Jenny et al. 2009. LabelMe Video: Building a video database with human annotations. In ICCV.
[3]
Pontus et al. 2012. Brat Rapid Annotation Tool. http://brat.nlplab.org/.
[4]
Russell et al. 2008. LabelMe: A Database and Web-Based Tool for Image Annotation. IJCV (2008).
[5]
Tsung-Yi et al. 2014. Microsoft COCO: Common Objects in Context. CoRR (2014).
[6]
Vondrick et al. 2013. Efficiently Scaling up Crowdsourced Video Annotation. IJCV (2013).
[7]
Vicol et al. 2018. MovieGraphs: Towards Understanding Human-Centric Situations from Videos. In CVPR.
[8]
Wagner et al. 2018. Applying Cooperative Machine Learning to Speed Up the Annotation of Social Signals in Large Multi-modal Corpora. arXiv:1802.02565 (2018).
[9]
H. Lausberg and H Sloetjes. 2009. Coding gestural behavior with the NEUROGES-ELAN system. BRM (2009).
[10]
Xiao Lin, <b>Chris Kim,</b> Timothy Meo, and Mohamed R. Amer. 2018. Learn, Generate, Rank: Generative Ranking of Motion Capture. In European Conference on Computer Vision.
[11]
Timothy Meo, <b>Chris Kim,</b> Aswin Raghavan, Alex Tozzo, David A. Salter, Amir Tamrakar, and Mohamed R. Amer. 2019. Aesop: A Visual Storytelling Platform for Conversational AI and Commonsense Grounding. AI Communications (2019).
[12]
Philip V. Ogren. 2006. Knowtator: A protégé plug-in for annotated corpus construction. In NAACL-HLT.
[13]
Rainer Simon. 2018. Annotorious - Image Annotation for the Web. https://annotorious.github.io/.

Cited By

View all
  • (2022)Tooling for Developing Data-Driven Applications: Overview and OutlookProceedings of Mensch und Computer 202210.1145/3543758.3543779(66-77)Online publication date: 4-Sep-2022

Index Terms

  1. A modular framework for collaborative multimodal annotation and visualization

      Recommendations

      Comments

      Please enable JavaScript to view thecomments powered by Disqus.

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      IUI '19 Companion: Companion Proceedings of the 24th International Conference on Intelligent User Interfaces
      March 2019
      173 pages
      ISBN:9781450366731
      DOI:10.1145/3308557
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 16 March 2019

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. HCI
      2. commonsense grounding
      3. conversational AI
      4. explainable AI
      5. language and vision
      6. multimodal annotation

      Qualifiers

      • Short-paper

      Conference

      IUI '19
      Sponsor:

      Acceptance Rates

      Overall Acceptance Rate 746 of 2,811 submissions, 27%

      Upcoming Conference

      IUI '25

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)7
      • Downloads (Last 6 weeks)1
      Reflects downloads up to 09 Jan 2025

      Other Metrics

      Citations

      Cited By

      View all
      • (2022)Tooling for Developing Data-Driven Applications: Overview and OutlookProceedings of Mensch und Computer 202210.1145/3543758.3543779(66-77)Online publication date: 4-Sep-2022

      View Options

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media