skip to main content
research-article

Seq2Set2Seq: A Two-stage Disentangled Method for Reply Keyword Generation in Social Media

Authors Info & Claims
Published:09 March 2024Publication History
Skip Abstract Section

Abstract

Social media produces large amounts of content every day. How to predict the potential influences of the contents from a social reply feedback perspective is a key issue that has not been explored. Thus, we propose a novel task named reply keyword prediction in social media, which aims to predict the keywords in the potential replies in as many aspects as possible. One prerequisite challenge is that the accessible social media datasets labeling such keywords remain absent. To solve this issue, we propose a new dataset,1 to study the reply keyword prediction in social media. This task could be seen as a single-turn dialogue keyword prediction for open-domain dialogue system. However, existing methods for dialogue keyword prediction cannot be adopted directly, which has two main drawbacks. First, they do not provide an explicit mechanism to model topic complementarity between keywords which is crucial in social media to controllably model all aspects of replies. Second, the collocations of keywords are not explicitly modeled, which also makes it less controllable to optimize for fine-grained prediction since the context information is much less than that in dialogue. To address these issues, we propose a two-stage disentangled framework, which can optimize the complementarity and collocation explicitly in a disentangled fashion. In the first stage, we use a sequence-to-set paradigm via multi-label prediction and determinantal point processes, to generate a set of keyword seeds satisfying the complementarity. In the second stage, we adopt a set-to-sequence paradigm via seq2seq model with the keyword seeds guidance from the set, to generate the more-fine-grained keywords with collocation. Experiments show that this method can generate not only a more diverse set of keywords but also more relevant and consistent keywords. Furthermore, the keywords obtained based on this method can achieve better reply generation results in the retrieval-based system than others.

REFERENCES

  1. [1] Wang Yue, Li Jing, Chan Hou Pong, King Irwin, Lyu Michael R., and Shi Shuming. 2019. Topic-aware neural keyphrase generation for social media language. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 25162526.Google ScholarGoogle ScholarCross RefCross Ref
  2. [2] Yuan Xingdi, Wang Tong, Meng Rui, Thaker Khushboo, Brusilovsky Peter, He Daqing, and Trischler Adam. 2018. One size does not fit all: Generating and evaluating variable number of keyphrases. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 79617975.Google ScholarGoogle Scholar
  3. [3] Huang Xiaoli, Xu Tongge, Jiao Lvan, Zu Yueran, and Zhang Youmin. 2021. Adaptive beam search decoding for discrete keyphrase generation. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 35. 1308213089.Google ScholarGoogle ScholarCross RefCross Ref
  4. [4] Ye Jiacheng, Gui Tao, Luo Yichao, Xu Yige, and Zhang Qi. 2021. One2Set: Generating diverse keyphrases as a set. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). Association for Computational Linguistics, Online, 45984608. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  5. [5] Tang Jianheng, Zhao Tiancheng, Xiong Chenyan, Liang Xiaodan, Xing Eric, and Hu Zhiting. 2019. Target-guided open-domain conversation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 56245634.Google ScholarGoogle ScholarCross RefCross Ref
  6. [6] Ishii Samuel Cahyawijaya, Zihan Liu, Genta Indra Winata, Andrea Madotto, Dan Su, Pascale Fung, Yan Xu, and Etsuko. 2022. Retrieval-free knowledge-grounded dialogue response generation with adapters. In Proceedings of the Second DialDoc Workshop on Document-grounded Dialogue and Conversational Question Answering. 93107.Google ScholarGoogle Scholar
  7. [7] Borodin Alexei. 2009. Determinantal point processes. arXiv preprint arXiv:0911.1153 (2009).Google ScholarGoogle Scholar
  8. [8] Chen Laming, Zhang Guoxin, and Zhou Hanning. 2018. Fast greedy MAP inference for determinantal point process to improve recommendation diversity. In Proceedings of the 32nd International Conference on Neural Information Processing Systems (NIPS’18). Curran Associates Inc., Red Hook, NY, USA, 56275638.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. [9] Elfeki M., Couprie C., Riviere M., and Elhoseiny M.. 2019. GDPP: Learning diverse generations using determinantal point processes. In Proceedings of International Conference on Machine Learning.Google ScholarGoogle Scholar
  10. [10] Shang Lifeng, Lu Zhengdong, and Li Hang. 2015. Neural responding machine for short-text conversation. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing. 15771586.Google ScholarGoogle ScholarCross RefCross Ref
  11. [11] Fleiss Joseph L.. 1971. Measuring nominal scale agreement among many raters. Psychological Bulletin 76, 5 (1971), 378.Google ScholarGoogle ScholarCross RefCross Ref
  12. [12] Sun Yu, Wang Shuohuan, Li Yukun, Feng Shikun, Chen Xuyi, Zhang Han, Tian Xin, Zhu Danxiang, Tian Hao, and Wu Hua. 2019. ERNIE: Enhanced representation through knowledge integration. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 14411451.Google ScholarGoogle Scholar
  13. [13] Jiang Di, Song Yuanfeng, Lian Rongzhong, Bao Siqi, Peng Jinhua, He Huang, Wu Hua, Zhang Chen, and Chen Lei. 2021. Familia: A configurable topic modeling framework for industrial text engineering. In Database Systems for Advanced Applications - 26th International Conference, DASFAA 2021, Taipei, Taiwan, April 11–14, 2021, Proceedings, Part III (LNCS), Vol. 12683. Springer, 516528. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. [14] Kim Yoon. 2014. Convolutional neural networks for sentence classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP’14). 17461751.Google ScholarGoogle ScholarCross RefCross Ref
  15. [15] Yi Le, Zhang Lei, Xu Xiuyuan, and Guo Jixiang. 2022. Multi-label softmax networks for pulmonary nodule classification using unbalanced and dependent categories. IEEE Transactions on Medical Imaging (2022).Google ScholarGoogle Scholar
  16. [16] Y. E. Lee and M. Lee. 2020. Decoding visual responses based on deep neural networks with ear-EEG signals. 8th International Winter Conference on Brain-Computer Interface (BCI’20), IEEE, 1–6.Google ScholarGoogle Scholar
  17. [17] Li Zheng, Wang Zijian, Tan Ming, Nallapati Ramesh, Bhatia Parminder, Arnold Andrew, Xiang Bing, and Roth Dan. 2022. DQ-BART: Efficient sequence-to-sequence model via joint distillation and quantization. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). Association for Computational Linguistics, Dublin, Ireland, 203211. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  18. [18] Bahdanau Dzmitry, Cho Kyunghyun, and Bengio Yoshua. 2014. Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473 (2014).Google ScholarGoogle Scholar
  19. [19] Vaswani Ashish, Shazeer Noam, Parmar Niki, Uszkoreit Jakob, Jones Llion, Gomez Aidan N., Kaiser Łukasz, and Polosukhin Illia. 2017. Attention is all you need. In Proceedings of Advances in Neural Information Processing Systems. 59986008.Google ScholarGoogle Scholar
  20. [20] Zhao Tiancheng, Zhao Ran, and Eskenazi Maxine. 2017. Learning discourse-level diversity for neural dialog models using conditional variational autoencoders. arXiv preprint arXiv:1703.10960 (2017).Google ScholarGoogle Scholar
  21. [21] Bousmalis Konstantinos, Trigeorgis George, Silberman Nathan, Krishnan Dilip, and Erhan Dumitru. 2016. Domain separation networks. In Proceedings of Advances in Neural Information Processing Systems. 343351.Google ScholarGoogle Scholar
  22. [22] Chen Jun, Zhang Xiaoming, Wu Yu, Yan Zhao, and Li Zhoujun. 2018. Keyphrase generation with correlation constraints. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. 40574066.Google ScholarGoogle ScholarCross RefCross Ref
  23. [23] See Abigail, Liu Peter J., and Manning Christopher D.. 2017. Get to the point: Summarization with pointer-generator networks. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics Volume 1 (Long Papers). 10731083.Google ScholarGoogle ScholarCross RefCross Ref
  24. [24] Xie Binbin, Wei Xiangpeng, Yang Baosong, Lin Huan, Xie Jun, Wang Xiaoli, Zhang Min, and Su Jinsong. 2022. WR-One2Set: Towards well-calibrated keyphrase generation. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Goldberg Yoav, Kozareva Zornitsa, and Zhang Yue (Eds.). Association for Computational Linguistics, Abu Dhabi, United Arab Emirates, 72837293. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  25. [25] Song Yan, Shi Shuming, Li Jing, and Zhang Haisong. 2018. Directional skip-gram: Explicitly distinguishing left and right context for word embeddings. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers). Association for Computational Linguistics, 175180. http://aclweb.org/anthology/N18-2028Google ScholarGoogle ScholarCross RefCross Ref
  26. [26] Santos Jorge M. and Embrechts Mark. 2009. On the use of the Adjusted Rand Index as a metric for evaluating supervised classification. In Proceedings of International Conference on Artificial Neural Networks. Springer, 175184.Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. [27] George Hripcsak and S. Adam Rothschild. 2005. Agreement, the F-Measure, and reliability in information retrieval. Journal of the American Medical Informatics Association JAMIA (2005).Google ScholarGoogle Scholar
  28. [28] Devlin Jacob, Chang Ming-Wei, Lee Kenton, and Toutanova Kristina. 2018. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). 41714186.Google ScholarGoogle Scholar
  29. [29] Xing Chen, Wu Wei, Wu Yu, Liu Jie, Huang Yalou, Zhou Ming, and Ma Wei-Ying. 2017. Topic aware neural response generation. In Proceedings of Thirty-First AAAI Conference on Artificial Intelligence.Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. [30] Vinyals O. and Le Q.. 2015. A neural conversational model. arXiv preprint arXiv:1506.05869 (2015).Google ScholarGoogle Scholar
  31. [31] Jiwei Li, Michel Galley, Chris Brockett, Jianfeng Gao, and Bill Dolan. 2016. A diversity-promoting objective function for neural conversation models. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Association for Computational Linguistics, San Diego, CA, 110–119.Google ScholarGoogle Scholar
  32. [32] Mou Lili, Song Yiping, Yan Rui, Li Ge, Zhang Lu, and Jin Zhi. 2016. Sequence to backward and forward sequences: A content-introducing approach to generative short-text conversation. In Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers. 33493358.Google ScholarGoogle Scholar
  33. [33] Serban Iulian Vlad, Klinger Tim, Tesauro Gerald, Talamadupula Kartik, Zhou Bowen, Bengio Yoshua, and Courville Aaron. 2017. Multiresolution recurrent neural networks: An application to dialogue response generation. In Proceedings of Thirty-First AAAI Conference on Artificial Intelligence.Google ScholarGoogle ScholarCross RefCross Ref
  34. [34] Zhu Qingfu, Zhang Weinan, Cui Lei, and Liu Ting. 2019. Order-sensitive keywords based response generation in open-domain conversational systems. ACM Transactions on Asian and Low-Resource Language Information Processing (TALLIP) 19, 2 (2019), 118.Google ScholarGoogle Scholar
  35. [35] Chan Hou Pong, Chen Wang, Wang Lu, and King Irwin. 2019. Neural keyphrase generation via reinforcement learning with adaptive rewards. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 21632174.Google ScholarGoogle ScholarCross RefCross Ref
  36. [36] Bang Namo, Lee Jeehyun, and Koo Myoung-Wan. 2023. Task-optimized adapters for an end-to-end task-oriented dialogue system. In Findings of the Association for Computational Linguistics: ACL 2023, Rogers Anna, Boyd-Graber Jordan, and Okazaki Naoaki (Eds.). Association for Computational Linguistics, Toronto, Canada, 73557369. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  37. [37] Luo Man, Fang Zhiyuan, Gokhale Tejas, Yang Yezhou, and Baral Chitta. 2023. End-to-end knowledge retrieval with multi-modal queries. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Rogers Anna, Boyd-Graber Jordan, and Okazaki Naoaki (Eds.). Association for Computational Linguistics, Toronto, Canada, 85738589. DOI:Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Seq2Set2Seq: A Two-stage Disentangled Method for Reply Keyword Generation in Social Media

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    • Published in

      cover image ACM Transactions on Asian and Low-Resource Language Information Processing
      ACM Transactions on Asian and Low-Resource Language Information Processing  Volume 23, Issue 3
      March 2024
      277 pages
      ISSN:2375-4699
      EISSN:2375-4702
      DOI:10.1145/3613569
      Issue’s Table of Contents

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 9 March 2024
      • Online AM: 5 February 2024
      • Accepted: 23 January 2024
      • Revised: 16 December 2023
      • Received: 26 July 2023
      Published in tallip Volume 23, Issue 3

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
    • Article Metrics

      • Downloads (Last 12 months)72
      • Downloads (Last 6 weeks)23

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    View Full Text