[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
research-article

-generAItor: Tree-in-the-loop Text Generation for Language Model Explainability and Adaptation

Published: 05 June 2024 Publication History

Abstract

Large language models (LLMs) are widely deployed in various downstream tasks, e.g., auto-completion, aided writing, or chat-based text generation. However, the considered output candidates of the underlying search algorithm are under-explored and under-explained. We tackle this shortcoming by proposing a tree-in-the-loop approach, where a visual representation of the beam search tree is the central component for analyzing, explaining, and adapting the generated outputs. To support these tasks, we present generAItor, a visual analytics technique, augmenting the central beam search tree with various task-specific widgets, providing targeted visualizations and interaction possibilities. Our approach allows interactions on multiple levels and offers an iterative pipeline that encompasses generating, exploring, and comparing output candidates, as well as fine-tuning the model based on adapted data. Our case study shows that our tool generates new insights in gender bias analysis beyond state-of-the-art template-based methods. Additionally, we demonstrate the applicability of our approach in a qualitative user study. Finally, we quantitatively evaluate the adaptability of the model to few samples, as occurring in text-generation use cases.

References

[1]
Davey Alba. 2022. OpenAI chatbot spits out biased musings, despite guardrails. Bloomberg. Retrieved from https://www.bloomberg.com/news/newsletters/2022-12-08/chatgpt-open-ai-s-chatbot-is-spitting-out-biased-sexist-results
[2]
Sarah Alnegheimish, Alicia Guo, and Yi Sun. 2022. Using natural sentence prompts for understanding biases in language models. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics, 2824–2830.
[3]
R. Arias-Hernandez, L. T. Kaastra, T. M. Green, and B. Fisher. 2011. Pair analytics: Capturing reasoning processes in collaborative visual analytics. In Proceedings of the Hawaii International Conference on System Sciences. IEEE.
[4]
Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2014. Neural machine translation by jointly learning to align and translate. arxiv:1409.0473
[5]
Yoshua Bengio, Réjean Ducharme, and Pascal Vincent. 2000. A neural probabilistic language model. Adv. Neural Inf. Process. Syst. 13 (2000).
[6]
Su Lin Blodgett, Solon Barocas, Hal Daumé III, and Hanna Wallach. 2020. Language (technology) is power: A critical survey of “Bias” in NLP. In Proceedings of the Association for Computational Linguistics. Association for Computational Linguistics, 5454–5476.
[7]
Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever and Dario Amodei. 2020. Language models are few-shot learners. arxiv:2005.14165
[8]
Aylin Caliskan, Joanna J. Bryson, and Arvind Narayanan. 2017. Semantics derived automatically from language corpora contain human-like biases. Science 356, 6334 (2017), 183–186.
[9]
Jose Camacho-Collados and Roberto Navigli. 2017. BabelDomains: Large-scale domain labeling of lexical resources. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics. Association for Computational Linguistics.
[10]
Ricardo Campos, Vítor Mangaravite, Arian Pasquali, Alípio Jorge, Célia Nunes, and Adam Jatowt. 2020. YAKE! Keyword extraction from single documents using multiple local features. Inf. Sci. 509 (2020), 257–289.
[11]
Catherine Chen, Kevin Lin, and Dan Klein. 2021. Constructing taxonomies from pretrained language models. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics.
[12]
Together Computer. 2023. RedPajama: An Open Source Recipe to Reproduce LLaMA Training Dataset. Retrieved from https://github.com/togethercomputer/RedPajama-Data
[13]
Simone Conia and Roberto Navigli. 2020. Conception: Multilingually-enhanced, human-readable concept vector representations. In Proceedings of the 28th International Conference on Computational Linguistics. International Committee on Computational Linguistics.
[14]
Norbert Corver and Henk van Riemsdijk. 2001. Semi-lexical Categories: The Function of Content Words and the Content of Function Words. De Gruyter Mouton, Berlin, New York.
[15]
Marina Danilevsky, Kun Qian, Ranit Aharonov, Yannis Katsis, Ban Kawas, and Prithviraj Sen. 2020. A survey of the state of explainable AI for natural language processing. In Proceedings of the 1st Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing. Association for Computational Linguistics, 447–459.
[16]
Sumanth Dathathri, Andrea Madotto, Janice Lan, Jane Hung, Eric Frank, Piero Molino, Jason Yosinski, and Rosanne Liu. 2019. Plug and play language models: A simple approach to controlled text generation. Retrieved from https://arxiv.org/abs/1912.02164
[17]
Deep NLP. 2023. Bias in NLP. Retrieved from https://github.com/cisnlp/bias-in-nlp
[18]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. BERT: Pre-training of deep bidirectional transformers for language understanding. arxiv:1810.04805
[19]
Wanyu Du, Zae Myung Kim, Vipul Raheja, Dhruv Kumar, and Dongyeop Kang. 2022. Read, revise, repeat: A system demonstration for human-in-the-loop iterative text revision. In Proceedings of the 1st Workshop on Intelligent and Interactive Writing Assistants. Association for Computational Linguistics.
[20]
M. El-Assady, W. Jentner, R. Kehlbeck, U. Schlegel, R. Sevastjanova, F. Sperrle, T. Spinner, and D. Keim. 2019. Towards XAI: Structuring the processes of explanations. In Proceedings of the ACM CHI Workshop: Human-centered Machine Learning Perspectives.
[21]
Mennatallah El-Assady, Rebecca Kehlbeck, Yannick Metz, Udo Schlegel, Rita Sevastjanova, Fabian Sperrle, and Thilo Spinner. 2022. Semantic color mapping: A pipeline for assigning meaningful colors to text. In Proceedings of the 4th IEEE Workshop on Visualization Guidelines in Research, Design, and Education.
[22]
Mennatallah El-Assady, Rita Sevastjanova, Daniel Keim, and Christopher Collins. 2018. ThreadReconstructor: Modeling reply-chains to untangle conversational text through visual analytics. Comput. Graph. Forum 37, 3 (2018), 351–365.
[23]
Kawin Ethayarajh. 2019. How contextual are contextualized word representations? Comparing the geometry of BERT, ELMo, and GPT-2 embeddings. In Proceedings of the Conference on Empirical Methods in Natural Language and the International Joint Conference on Natural Language Processing. ACL, 55–65.
[24]
Ismael Garrido-Muñoz, Arturo Montejo-Ráez, Fernando Martínez-Santiago, and L. Alfonso Ureña-López. 2021. A survey on bias in deep NLP. Appl. Sci. 11, 7 (2021), 3184.
[25]
Albert Gatt and Emiel Krahmer. 2018. Survey of the state of the art in natural language generation: Core tasks, applications and evaluation. J. Artif. Intell. Res. 61 (2018), 65–170.
[26]
Sebastian Gehrmann, Hendrik Strobelt, Robert Kruger, Hanspeter Pfister, and Alexander M. Rush. 2019. Visual interaction with deep learning models through collaborative semantic inference. IEEE Trans. Visualiz. Comput. Graph. (2019), 1–1.
[27]
Jochen Hartmann, Mark Heitmann, Christina Schamp, and Oded Netzer. 2021. The power of brand selfies. J. Market. Res. 58, 6 (2021).
[28]
Xingwei He. 2021. Parallel refinements for lexically constrained text generation with BART. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics.
[29]
Jeremy Howard and Sebastian Ruder. 2018. Universal language model fine-tuning for text classification. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, 328–339.
[30]
Zhiting Hu, Zichao Yang, Xiaodan Liang, Ruslan Salakhutdinov, and Eric P. Xing. 2017. Toward controlled generation of text. In Proceedings of the 34th International Conference on Machine Learning (Proceedings of Machine Learning Research), Doina Precup and Yee Whye Teh (Eds.), Vol. 70. PMLR, 1587–1596.
[31]
Xinyu Hua and Lu Wang. 2020. PAIR: Planning and iterative refinement in pre-trained transformers for long text generation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP’20). Association for Computational Linguistics.
[32]
Jiaxin Huang, Yiqing Xie, Yu Meng, Yunyi Zhang, and Jiawei Han. 2020. CoRel: Seed-guided topical taxonomy construction by concept learning and relation transferring. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM.
[33]
Fred Jelinek, Robert L. Mercer, Lalit R. Bahl, and James K. Baker. 1977. Perplexity—A measure of the difficulty of speech recognition tasks. J. Acoustic. Soc. Amer. 62, S1 (1977), S63–S63.
[34]
Ziwei Ji, Nayeon Lee, Rita Frieske, Tiezheng Yu, Dan Su, Yan Xu, Etsuko Ishii, Ye Jin Bang, Andrea Madotto, and Pascale Fung. 2023. Survey of hallucination in natural language generation. Comput. Surv. 55, 12 (2023), 1–38.
[35]
Minhao Jiang, Xiangchen Song, Jieyu Zhang, and Jiawei Han. 2022. TaxoEnrich: Self-supervised taxonomy completion via structure-semantic representations. In Proceedings of the ACM Web Conference. ACM.
[36]
Jeff Johnson, Matthijs Douze, and Hervé Jégou. 2019. Billion-scale similarity search with GPUs. IEEE Trans. Big Data 7, 3 (2019), 535–547.
[37]
Aikaterini-Lida Kalouli, Rita Sevastjanova, Christin Beck, and Maribel Romero. 2022. Negation, coordination, and quantifiers in contextualized language models. In Proceedings of the 29th International Conference on Computational Linguistics.International Committee on Computational Linguistics, 3074–3085.
[38]
Rebecca Kehlbeck, Rita Sevastjanova, Thilo Spinner, Tobias Stähle, and Mennatallah El-Assady. 2021. Demystifying the embedding space of language models. In Proceedings of the Workshop on Visualization for AI Explainability (VISxAI’21). Retrieved from https://bert-vs-gpt2.dbvis.de/
[39]
Anne Lauscher, Tobias Lueken, and Goran Glavaš. 2021. Sustainable modular debiasing of language models. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP. Association for Computational Linguistics, 4782–4797.
[40]
Yann LeCun. 2023. Do Language Models Need Sensory Grounding for Meaning and Understanding? Retrieved from https://drive.google.com/file/d/1BU5bV3X5w65DwSMapKcsr0ZvrMRU_Nbi
[41]
Jaesong Lee, Joong-Hwi Shin, and Jun-Seok Kim. 2017. Interactive visualization and manipulation of attention-based neural machine translation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing: System Demonstrations. Association for Computational Linguistics, 121–126.
[42]
Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Veselin Stoyanov, and Luke Zettlemoyer. 2020. BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, 7871–7880.
[43]
Alexander Lex, Nils Gehlenborg, Hendrik Strobelt, Romain Vuillemot, and Hanspeter Pfister. 2014. UpSet: Visualization of intersecting sets. IEEE Trans. Visualiz. Comput. Graph. 20, 12 (2014), 1983–1992.
[44]
Junyi Li, Tianyi Tang, Wayne Xin Zhao, and Ji-Rong Wen. 2021. Pretrained language model for text generation: A survey. In Proceedings of the 30th International Joint Conference on Artificial Intelligence. International Joint Conference on Artificial Intelligence Organization.
[45]
Zhuliu Li, Yiming Wang, Xiao Yan, Weizhi Meng, Yanen Li, and Jaewon Yang. 2022. TaxoTrans. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. ACM.
[46]
Paul Pu Liang, Chiyu Wu, Louis-Philippe Morency, and Ruslan Salakhutdinov. 2021. Towards understanding and mitigating social biases in language models. In Proceedings of the International Conference on Machine Learning. PMLR, 6565–6576.
[47]
Ilya Loshchilov and Frank Hutter. 2017. Fixing weight decay regularization in adam. CoRR abs/1711.05101 (2017).
[48]
Kaiji Lu, Piotr Mardziel, Fangjing Wu, Preetam Amancharla, and Anupam Datta. 2020. Gender bias in neural natural language processing. Logic, Language, and Security: Essays Dedicated to Andre Scedrov on the Occasion of His 65th Birthday (2020), Springer International Publishing, 189–202.
[49]
Andrew L. Maas, Raymond E. Daly, Peter T. Pham, Dan Huang, Andrew Y. Ng, and Christopher Potts. 2011. Learning word vectors for sentiment analysis. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics, 142–150.
[50]
Leland McInnes, John Healy, Nathaniel Saul, and Lukas Grossberger. 2018. UMAP: Uniform manifold approximation and projection. J. Open Source Softw. 3, 29 (2018), 861.
[51]
Ninareh Mehrabi, Fred Morstatter, Nripsuta Saxena, Kristina Lerman, and Aram Galstyan. 2021. A survey on bias and fairness in machine learning. Comput. Surv. 54, 6 (2021), 1–35.
[52]
Cade Metz. 2022. The new chatbots could change the world. Can you trust them? New York Times. Retrieved from https://www.nytimes.com/2022/12/10/technology/ai-chat-bot-chatgpt.html
[53]
Swaroop Mishra, Daniel Khashabi, Chitta Baral, and Hannaneh Hajishirzi. 2022. Cross-task generalization via natural language crowdsourcing instructions. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics.
[54]
Andrea Moro, Alessandro Raganato, and Roberto Navigli. 2014. Entity linking meets word sense disambiguation: A unified approach. Trans. Assoc. Computat. Ling. 2 (2014), 231–244.
[55]
Moin Nadeem, Anna Bethke, and Siva Reddy. 2021. StereoSet: Measuring stereotypical bias in pretrained language models. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing. Association for Computational Linguistics, 5356–5371.
[56]
Roberto Navigli and Simone Paolo Ponzetto. 2012. BabelNet: The automatic construction, evaluation and application of a wide-coverage multilingual semantic network. Artif. Intell. 193 (2012), 217–250.
[57]
OpenAI. 2023. GPT-4 Technical Report. (2023). arxiv:2303.08774
[58]
Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, John Schulman, Jacob Hilton, Fraser Kelton, Luke Miller, Maddie Simens, Amanda Askell, Peter Welinder, Paul F. Christiano, Jan Leike, and Ryan Lowe. 2022. Training language models to follow instructions with human feedback. In Advances in Neural Information Processing Systems. S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh (Eds.), Vol. 35. Curran Associates, Inc., 27730–27744.
[59]
Vishakh Padmakumar and He He. 2022. Machine-in-the-loop rewriting for creative image captioning. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics.
[60]
Romain Paulus, Caiming Xiong, and Richard Socher. 2017. A deep reinforced model for abstractive summarization. CoRR abs/1705.04304 (2017).
[61]
Lianhui Qin, Vered Shwartz, Peter West, Chandra Bhagavatula, Jena D. Hwang, Ronan Le Bras, Antoine Bosselut, and Yejin Choi. 2020. Back to the future: Unsupervised backprop-based decoding for counterfactual and abductive commonsense reasoning. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP’20). Association for Computational Linguistics.
[62]
Alec Radford, Jeffrey Wu, Dario Amodei, Daniela Amodei, Jack Clark, Miles Brundage, and Ilya Sutskever. 2019a. Better Language Models and Their Implications. Retrieved from https://openai.com/blog/better-language-models/
[63]
Alec Radford, Jeff Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019b. Language models are unsupervised multitask learners. OpenAI blog 1, 8 (2019a), 9 Pages.
[64]
Emily Reif, Ann Yuan, Martin Wattenberg, Fernanda B. Viegas, Andy Coenen, Adam Pearce, and Been Kim. 2019. Visualizing and measuring the geometry of BERT. In Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. d’Alché Buc, E. Fox, and R. Garnett (Eds.). Curran Associates, Inc., 8594–8603.
[65]
Anna Rogers, Olga Kovaleva, and Anna Rumshisky. 2020. A primer in BERTology: What we know about how BERT works. Trans. Assoc. Computat. Ling. 8 (2020), 842–866.
[66]
Kevin Roose. 2023. How chatbots and large language models, or LLMs, actually work. New York Times. Retrieved from https://www.nytimes.com/2023/03/28/technology/ai-chatbots-chatgpt-bing-bard-llm.html
[67]
David E. Rumelhart, Geoffrey E. Hinton, and Ronald J. Williams. 1986. Learning representations by back-propagating errors. Cahiers De La Revue De Theologie Et De Philosophie 323, 6088 (1986), 533–536.
[68]
Teven Le Scao, Angela Fan, Christopher Akiki, Ellie Pavlick, Suzana Ilić, Daniel Hesslow, Roman Castagné, Alexandra Sasha Luccioni, François Yvon, Matthias Gallé, et al. 2023. BLOOM: A 176B-Parameter Open-Access Multilingual Language Model. (2023). arxiv:2211.05100
[69]
Bianca Scarlini, Tommaso Pasini, and Roberto Navigli. 2020. With more contexts comes better performance: Contextualized sense embeddings for all-round word sense disambiguation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics.
[70]
Rita Sevastjanova and Mennatallah El-Assady. 2022. Beware the rationalization trap! When language model explainability diverges from our mental models of language. In Proceedings of the Communication in Human-AI Interaction Workshop at IJCAI-ECAI. abs/2207.06897 (2022).
[71]
Rita Sevastjanova, Aikaterini-Lida Kalouli, Christin Beck, Hanna Hauptmann, and Mennatallah El-Assady. 2022. LMFingerprints: Visual explanations of language model embedding spaces through layerwise contextualization scores. Comput. Graph. Forum 41, 3 (2022), 295–307.
[72]
Thilo Spinner, Rebecca Kehlbeck, Rita Sevastjanova, Tobias Stähle, Daniel A. Keim, Oliver Deussen, Andreas Spitz, and Mennatallah El-Assady. 2023. Revealing the Unwritten: Visual Investigation of Beam Search Trees to Address Language Model Prompting Challenges. arxiv:2310.11252 (2023).
[73]
Thilo Spinner, Udo Schlegel, Hanna Schafer, and Mennatallah El-Assady. 2020. explAIner: A visual analytics framework for interactive and explainable machine learning. IEEE Trans. Visualiz. Comput. Graph. 26, 1 (2020).
[74]
Martin Steiger, J. Bernard, Simon Thum, Sebastian Mittelstädt, Marco Hutter, Daniel A. Keim, and Jörn Kohlhammer. 2015. Explorative analysis of 2D color maps. In Proceedings of the Computer Graphics, Visualization & Vision Conference (WSCG’15).
[75]
Robert J. Sternberg and Karin Sternberg. 2016. Cognitive Psychology. Nelson Education.
[76]
Hendrik Strobelt, Sebastian Gehrmann, Michael Behrisch, Adam Perer, Hanspeter Pfister, and Alexander M. Rush. 2018. Seq2Seq-Vis: A visual debugging tool for sequence-to-sequence models. IEEE Trans. Visualiz. Comput. Graph. 25, 1 (2018), 353–363.
[77]
Hendrik Strobelt, Jambay Kinley, Robert Krueger, Johanna Beyer, Hanspeter Pfister, and Alexander M. Rush. 2022. GenNI: Human-AI collaboration for data-backed text generation. IEEE Trans. Visualiz. Comput. Graph. 28, 1 (2022), 1106–1116.
[78]
Yanchao Tan, Carl Yang, Xiangyu Wei, Chaochao Chen, Longfei Li, and Xiaolin Zheng. 2022. Enhancing recommendation with automated tag taxonomy construction in hyperbolic space. In Proceedings of the IEEE 38th International Conference on Data Engineering (ICDE’22). IEEE.
[79]
A. J. Teuling, R. Stöckli, and S. I. Seneviratne. 2010. Bivariate colour maps for visualizing climate data. Int. J. Climatol. 31, 9 (2010), 1408–1412.
[80]
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. arxiv:1706.03762 (2017).
[81]
Patrick von Platen. 2020. How to Generate Text: Using Different Decoding Methods for Language Generation with Transformers. Retrieved from https://huggingface.co/blog/how-to-generate
[82]
Gregor Wiedemann, Steffen Remus, Avi Chawla, and Chris Biemann. 2019. Does BERT make any sense? Interpretable word sense disambiguation with contextualized embeddings. In Proceedings of KONVENS.
[83]
Adina Williams, Nikita Nangia, and Samuel Bowman. 2018. A broad-coverage challenge corpus for sentence understanding through inference. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics, 1112–1122.
[84]
Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Rémi Louf, Morgan Funtowicz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander M. Rush. 2020. Transformers: State-of-the-art natural language processing. In Proceedings of the Conference on Empirical Methods in Natural Language Processing: System Demonstrations. Association for Computational Linguistics, 38–45.
[85]
Yuejia Xiang, Ziheng Zhang, Jiaoyan Chen, Xi Chen, Zhenxi Lin, and Yefeng Zheng. 2021. OntoEA: Ontology-guided entity alignment via joint knowledge graph embedding. In Proceedings of the Findings of the Association for Computational Linguistics: ACL-IJCNLP. Association for Computational Linguistics.
[86]
Hongyuan Xu, Yunong Chen, Zichen Liu, Yanlong Wen, and Xiaojie Yuan. 2022. TaxoPrompt: A prompt-based generation method with taxonomic context for self-supervised taxonomy expansion. In Proceedings of the 31st International Joint Conference on Artificial Intelligence. International Joint Conference on Artificial Intelligence Organization.
[87]
Wenhao Yu, Chenguang Zhu, Zaitang Li, Zhiting Hu, Qingyun Wang, Heng Ji, and Meng Jiang. 2022. A survey of knowledge-enhanced text generation. Comput. Surv. 54, 11s (2022), 1–38.
[88]
Ann Yuan, Andy Coenen, Emily Reif, and Daphne Ippolito. 2022. Wordcraft: Story writing with large language models. In Proceedings of the 27th International Conference on Intelligent User Interfaces. ACM.
[89]
Chao Zhang, Fangbo Tao, Xiusi Chen, Jiaming Shen, Meng Jiang, Brian Sadler, Michelle Vanni, and Jiawei Han. 2018. TaxoGen. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM.
[90]
Hanqing Zhang, Haolin Song, Shaoyu Li, Ming Zhou, and Dawei Song. 2022. A survey of controllable text generation using transformer-based pre-trained language models. arxiv:2201.05337 (2022).
[91]
Haiyan Zhao, Hanjie Chen, Fan Yang, Ninghao Liu, Huiqi Deng, Hengyi Cai, Shuaiqiang Wang, Dawei Yin, and Mengnan Du. 2024. Explainability for large language models: A survey. ACM Trans. Intell. Syst. Technol. 2, 15 (2024).

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Transactions on Interactive Intelligent Systems
ACM Transactions on Interactive Intelligent Systems  Volume 14, Issue 2
June 2024
201 pages
EISSN:2160-6463
DOI:10.1145/3613555
Issue’s Table of Contents

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 05 June 2024
Online AM: 14 March 2024
Accepted: 30 January 2024
Revised: 26 January 2024
Received: 18 July 2023
Published in TIIS Volume 14, Issue 2

Check for updates

Author Tags

  1. Large language models
  2. beam search tree
  3. natural language generation
  4. explainability
  5. language transformers
  6. visual analytics

Qualifiers

  • Research-article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 429
    Total Downloads
  • Downloads (Last 12 months)429
  • Downloads (Last 6 weeks)41
Reflects downloads up to 10 Dec 2024

Other Metrics

Citations

View Options

Login options

Full Access

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Full Text

View this article in Full Text.

Full Text

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media