Argumentative Stance Prediction: An Exploratory Study on Multimodality and Few-Shot Learning Permalink
Published in The 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2023
To advance argumentative stance prediction as a multimodal problem, the First Shared Task in Multimodal Argument Mining hosted stance prediction in crucial social topics of gun control and abortion. Our exploratory study attempts to evaluate the necessity of images for stance prediction in tweets and compare outof-the-box text-based large-language model(LLM) in few-shot settings against fine-tuned unimodal and multimodal models. Our work suggests an ensemble of fine-tuned text-based language models (0.817 F1-score) outperforms both the multimodal (0.677 F1-score) and textbased few-shot prediction using a recent stateof-the-art LLM (0.550 F1-score). In addition to the differences in performance, our findings suggest that the multimodal models tend to perform better when image content is summarized as natural language over their native pixel structure and, using in-context examples improves few-shot performance of LLMs.Code Poster
Recommended citation: @inproceedings{sharma-etal-2023-argumentative, title = "Argumentative Stance Prediction: An Exploratory Study on Multimodality and Few-Shot Learning", author = "Sharma, Arushi and Gupta, Abhibha and Bilalpur, Maneesh", editor = "Alshomary, Milad and Chen, Chung-Chi and Muresan, Smaranda and Park, Joonsuk and Romberg, Julia", booktitle = "Proceedings of the 10th Workshop on Argument Mining", month = dec, year = "2023", address = "Singapore", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2023.argmining-1.18", doi = "10.18653/v1/2023.argmining-1.18", pages = "167--174" }