Implementation of Autoregressive Language Models for Generation of Seasonal Fixed-form Haiku in Japanese

Authors

  • Kodai Hirata Hokkaido University
  • Soichro Yokoyama Hokkaido University
  • Tomohisa Yamashita Hokkaido University
  • Hidenori Kawamura Hokkaido University

DOI:

https://doi.org/10.52731/liir.v003.075

Keywords:

haiku generation, natural language generation, human evaluation, language model

Abstract

This paper describes the implementation of an artificial intelligence haiku generator. We trained language models using existing haiku and literary studies, evaluated model performance using automatically computable evaluation indices such as perplexity, and subjectively evaluated the generated haiku by using a questionnaire.
The main contributions of this paper are as follows. First, the effectiveness of a series of model evaluation processes, including automatically calculable evaluation indices and the results of subjective evaluations using questionnaires, is investigated. These processes are effective in the development of haiku generation models. Second, high-quality haiku generation is achieved using high-performance language models such as GPT-2 and BART.
The results of the questionnaire survey revealed that it is possible to generate sensible haiku comparable to those written by humans.
The insight gained from this study is applicable to other generative tasks.

References

Xianchao Wu, Momo Klyen, Kazushige Ito, and Zhan Chen. Haiku generation using

deep neural networks. In The Association for Natural Language Processing, 2017.

Sepp Hochreiter and Jürgen Schmidhuber. Long Short-Term Memory. Neural

Computation, 9(8):1735–1780, 11 1997.

A. Radford, Jeffrey Wu, R. Child, David Luan, Dario Amodei, and Ilya Sutskever. Language

models are unsupervised multitask learners. In OpenAI Blog, 2019.

Mika Hämäläinen, Khalid Alnajjar, and Thierry Poibeau. Modern French Poetry Generation

with RoBERTa and GPT-2. In 13th International Conference on Computational Creativity

(ICCC) 2022, Bolzano, Italy, June 2022. ICCC.

Xingxing Zhang and Mirella Lapata. Chinese poetry generation with recurrent neural networks. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language

Processing (EMNLP), pages 670–680, Doha, Qatar, October 2014. Association for Computational Linguistics.

Vaswani et al. Attention is all you need. In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach,

R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017.

Naoko Tosa, Hideto Obara, and Michihiko Minoh. Hitch haiku: An interactive supporting

system for composing haiku poem. In Scott M. Stevens and Shirley J. Saldamarco, editors,

Entertainment Computing - ICEC 2008, pages 209–216, Berlin, Heidelberg, 2009. Springer

Berlin Heidelberg.

Rafal Rzepka and Kenji Araki. Haiku generator that reads blogs and illustrates them with

sounds and images. In Proceedings of the 24th International Conference on Artificial Intelligence, IJCAI’15, page 2496–2502. AAAI Press, 2015.

Lantao Yu, Weinan Zhang, Jun Wang, and Yong Yu. Seqgan: Sequence generative adversarial nets with policy gradient. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, AAAI’17, page 2852–2858. AAAI Press, 2017.

Konishi et al. Generation of haiku preferable for ordinary people by seqgan. In Information

Processing Society of Japan, Kansai Branch Section Convention, 09 2017.

Mika Hämäläinen and Khalid Alnajjar. Human evaluation of creative nlg systems: An interdisciplinary survey on recent papers. In Proceedings of the 1st Workshop on Natural

Language Generation, Evaluation, and Metrics (GEM 2021), page 84–95, United States,

Workshop on Natural Language Generation, Evaluation, and Metrics, GEM Workshop.

Taku Kudo, Kaoru Yamamoto, and Yuji Matsumoto. Applying conditional random fields to

Japanese morphological analysis. In Proceedings of the 2004 Conference on Empirical

Methods in Natural Language Processing, pages 230–237, Barcelona, Spain, July 2004.

Association for Computational Linguistics.

Stephen Merity, Nitish Shirish Keskar, and Richard Socher. Regularizing and optimizing

LSTM language models. 6th International Conference on Learning Representations, ICLR

- Conference Track Proceedings, 2018.

Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed,

Omer Levy, Veselin Stoyanov, and Luke Zettlemoyer. BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational

Linguistics, pages 7871–7880, Online, July 2020. Association for Computational Linguistics.

Wolf et al. Transformers: State-of-the-art natural language processing. In Proceedings of the

Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 38–45, Online, October 2020. Association for Computational Linguistics.

Downloads

Published

2023-02-17