cPAPERS: A Dataset of Situated and Multimodal Interactive Conversations in Scientific Papers (2024)

Anirudh Sundar  Jin Xu  William Gay  Christopher Richardson  Larry Heck
AI Virtual Assistant Lab
Georgia Institute of Technology
Equal Contribution
{asundar34, jxu81, wgay7, crichardson8, larryheck}@gatech.edu

Abstract

An emerging area of research in situated and multimodal interactive conversations (SIMMC) includes interactions in scientific papers. Since scientific papers are primarily composed of text, equations, figures, and tables, SIMMC methods must be developed specifically for each component to support the depth of inquiry and interactions required by research scientists. This work introduces Conversational Papers (cPAPERS), a dataset of conversational question-answer pairs from reviews of academic papers grounded in these paper components and their associated references from scientific documents available on arXiv. We present a data collection strategy to collect these question-answer pairs from OpenReview and associate them with contextual information from LaTeXsource files. Additionally, we present a series of baseline approaches utilizing Large Language Models (LLMs) in both zero-shot and fine-tuned configurations to address the cPAPERS dataset.

1 Introduction

Developing conversational assistants capable of situated and multimodal interactive conversations (SIMMC) over structured knowledge sources remains an open problem [32, 45, 26].

An emerging area of research within this domain is conversational interactions over scientific documents [39]. The number of scientific articles has grown dramatically over the past decade, making it difficult for scientists to find, read, understand, and connect advancements published by their fellow researchers. Scientists in the field of biomedicine, for example, publish over 1 million articles per year or a new article every 2 minutes on average [27]. Developing methods to understand scientific documents and assist researchers is an important problem, especially for the Natural Language Processing (NLP) community. Scientific documents present an interesting challenge in SIMMC-based methods since the content is frequently multimodal [5].

Besides textual paragraphs, researchers rely on various modalities to describe research methods. Figures convey information about concepts developed throughout the paper. Generally, figures represent many different types of information including most commonly graphical and image information. For example, figures explain model architectures and pipelines, summarize experimental results in graphical plots, and represent complex information such as gradient learning surfaces and feature maps. In addition, figures can be used to show various stages of image processing in computer vision and image generation papers.

Equations are crucial for grasping mathematical concepts in scientific texts but can be challenging to interpret. Equations frequently rely on notation introduced elsewhere in the document, often requiring readers to draw from the entire paper to understand the formulation. In addition, equations are typically related to and in many cases help the reader understand other paper components such as figures and tables (and vice-versa).

Finally, tables typically have structures that convey semantic information. These include the text in the row and column headings, and text and often links in the table cells. These structures summarize multiple concepts (e.g., experimental results) in a human-readable form. Therefore, developing SIMMC methods for scientific papers must include a semantic understanding of these structures and table content.

To advance the development of conversational assistants capable of SIMMC in scientific papers, this paper introduces Conversational Papers (cPAPERS111The dataset is publicly available at https://huggingface.co/datasets/avalab/cPAPERS. The collection process, baseline models, and code will be released with the camera-ready version under GNU Public License.), a dataset of conversations in English situated in equations (cPAPERS-EQNS), figures (cPAPERS-FIGS), and tabular information (cPAPERS-TBLS) for scientific texts. Question-answer pairs are sourced from reviews and rebuttals from OpenReview222https://openreview.net/. Textual grounding to answer the questions and answers are sourced from the corresponding scientific articles hosted on arXiv333https://arxiv.org, an open-access repository of academic preprints.

The contributions of this work include:

  • Introduction of the Conversational Papers (cPAPERS) dataset including three splits (EQNS, FIGS, TBLS)

  • Development of a novel and scalable approach for collecting question-answer pairs from OpenReview and linking them with relevant contextual information from open access tex sources on arXiv

  • Creation of baseline LLM approaches that utilize weakly grounded multimodal context for dialogue responses

The rest of the paper is organized as follows: Section 2 discusses related work in language modeling and question-answering for equations, tables, and figures. Section 3 provides a description of cPAPERS and details the dataset collection process. Next Sections 4 describe baseline approaches to address this datasetwith a series of experiments utilizing zero-shot prompting and parameter-efficient fine-tuning. Section 5 discusses the results, while Sections 6 and 7 address the conclusion and limitations, respectively.

cPAPERS: A Dataset of Situated and Multimodal Interactive Conversations in Scientific Papers (1)

2 Related Work

2.1 Language Modeling for Equations

The challenge of modeling mathematical equations has become an active research area in Natural Language Processing. [10] proposes an encoder-decoder approach to generate equations from math word problems. [51] presents an approach for the dual problem of generating math word problems consistent with equations. Prior work has also proposed learning representations from equations using the Transformer encoder [37], Tree-based encoders [52], and RNNs [54]. More recent work utilizing generative architectures includes MathGPT [42], an auto-regressive model based on GPT-2 [38] for various language+equation tasks, and Nougat [5], a Visual Transformer to parse academic documents to markup language. In contrast, cPAPERS is a dataset that addresses both grounded conversational question-answering and the modeling of mathematical equations. Answering the questions accurately requires understanding the expressions in the equations while also formulating an accurate response that addresses the task succinctly.

2.2 Question-Answering over Figures

Early datasets on question-answering over figures include VQA [1], Visual7W [61], VisDial [12], ManyModalQA [20], and OK-VQA [31, 40]. More recent datasets addressing the problem of open-domain conversations include IGC [33], MOD [16], and Image-Chat [43]. However, the images in these datasets are collected from MS-COCO [29] or YFCC100M [49], where visual content targets commonly seen everyday objects as opposed to scientific documents where images target specific information relevant to the explanation of a concept.

Recent work has addressed some of the challenges associated with multimodal image+text tasks situated in scientific documents. [48] presents a dataset of charts from scientific papers and associated natural language captions summarizing the information present in the chart. [18] develops a method to link labels with relevant images in patents. [11] introduces a dataset of 150 computer science papers, ground truth labels for the locations of figures, tables, and captions, and an approach to automatically extract this information from PDFs.SciCap [24] is a much larger dataset of 400,000 figures, their captions, and associated textual references from various scientific papers. The dataset is collected by scraping scientific preprints from arXiv. For each figure, they provide the associated caption and all paragraphs that mention the figure.

While these datasets address tasks related to multimodal content in scientific documents, they are not conversational in nature.In contrast, cPAPERS addresses the shortcomings of prior datasets as a multimodal, conversational dataset grounded in scientific documents. The questions and answers address specific contextual visual information grounded in scientific documents.

2.3 Tabular Question-Answering

Tabular question-answering addresses the problem of extractive QA grounded in the information contained in specific cells of a table.Prior tabular QA datasets include those collected from Wikipedia, such as WikiTableQuestions [36], ManyModalQA [21], TaBERT [55], NQ-Tables [22], FEVEROUS [3], FeTaQA [35], HybriDialogue [34], and HiTab [9]. Other tabular datasets are constructed from financial reports including TAT-QA [60], FinQA [8], MultiHiertt [59], or arXiv iTBLS [46].

Proposed approaches to address the tabular QA task include architectures based off of the Transformer encoder [55, 23, 7, 15, 30, 19, 53], decoder [17, 2, 56, 25, 57, 44], or both (encoder-decoder) [34, 13, 47, 41, 46].

In contrast to prior tabular datasets, cPAPERS-TBLS presents a new source of grounded questions and answers and to the best of our knowledge, is the first to be situated in tables from OpenReview. Additionally, the question-answer pairs are not factoids. Rather, they are conversational in nature.

EquationTableFigure
traindevtesttraindevtesttraindevtest
# Unique Papers672286335715285302761275313
# QA Pairs9933363949323273421052297357
# Tokens (average)
traindevtesttraindevtesttraindevtest
Question252525242226232324
Answer9210290867981838881
Contexts10,23211,85112,2882,9812,7462,610433400431
References7,3238,4139,5171,7571,6451,427366375323
Neighboring Contexts1,1441,1531,0849941,043925---
Neighboring References1,0009471,152736657588---

3 The cPAPERS Dataset

3.1 Dataset Description

Conversational Papers (cPAPERS) is a dataset of conversations in English situated in scientific texts. cPAPERS consists of question-answer pairs pertaining to figures (cPAPERS-FIGS), equations (cPAPERS-EQNS), or tabular information (cPAPERS-TBLS) from scientific papers. These pairs are sourced from the official reviews and rebuttals of The Conference and Workshop on Neural Information Processing Systems (NeurIPS) and The International Conference on Learning Representations (ICLR) between 2020 and 2023, as available on OpenReview. cPAPERS comprises of 1723 question-answer pairs over equations, 1601 pairs over tables, and 1706 pairs over figures, totaling 5030 question-answer pairs spanning 2350 unique scientific papers. A detailed breakdown of the dataset is provided in Table 1.

Section 3.2 details the steps of collecting cPAPERS from official reviews on OpenReview, extracting question-answer pairs, and associating them with LaTeXsource files from arXiv.

3.2 Dataset Collection

3.2.1 Step 1 - Extract Official Review and Comments

We leverage the OpenReview API to access official reviews and comments for each paper submission. These reviews, authored by conference reviewers, typically include a paper summary, strengths and weaknesses, specific questions, and limitations. Authors respond to these critiques in the comments, providing answers that address the reviewers’ questions. This step involves downloading the official reviews and comments for all submissions.

3.2.2 Step 2 - Download LaTeXfiles

For each paper recorded in Step 1, we initiate a call to the arXiv API to verify the existence of the paper. If the paper is found, we proceed to download all associated LaTeXsource files. If a paper cannot be located using the arXiv API, we exclude this submission from further processing.

3.2.3 Step 3 - Regex Filtering

As mentioned in Step 1, review-rebuttal pairs contain a summary of the paper and additional information that may not be related to equations, tables, or figures. Therefore, we use regular expressions to identify mentions of equations, tables, and figures within each review and comment pair. The following regular expressions are employed for this purpose:

Equation: re.compile(r’\b(?:Equation|Eq\.?)\s*\(?\s*\d+\)?\b’, re.IGNORECASE)
Table: re.compile(r’\bTable\s*\d+|\btables\b’,re.IGNORECASE)
Figure: re.compile(r’\b(?:Figure|fig)\.?\s*\d+\b’, re.IGNORECASE)

These expressions efficiently determine whether equations, tables, and figures are mentioned in either the review or the rebuttal, resulting in a separate list of review-rebuttal pairs each containing explicit references to an equation, table, or figure in the paper. It is possible for a review to contain references to multiple modalities.

3.2.4 Step 4 - Question-Answer Extraction

To extract question-answer pairs from the regex-filtered reviews and rebuttals, an LLM that is instruction fine-tuned to align with human preferences is prompted in-context with a single example. For this process, we use Llama-2-70b-chat-hf [50].

The prompt is as follows:

‘Given this example Content, Question, and Answer, look through the Summary and Comments that are dedicated to the paper and extract Question and Answer pairs that specifically target a particular Figure/Equation/Table in the content. Questions may be under sections labeled similarly to “Weaknesses" or “Questions” or “Review" or “Clarity, Quality, Novelty And Reproducibility" in the comments, and Answers may be under sections labeled “Response to Reviewer" or “Rebuttal" or “Comments".’

System Prompt: “You are a helpful question answer extracting assistant. Our goal is to extract Question and Answer pairs pertaining to a specific a Figure/Equation/Table in the content. Your response should be in the format of [Question] <question> [Answer] <answer> Do not add any other unnecessary content in your response."

Post-processing + JSON reformatting.GPT 3.5 [6] is used to reformat the question-answer pairs as JSON. The following system prompt is used.

Prompt: You are a helpful assistant. Please find the Question Answer pairs, and format it as a json ({“question_answers": [{ “question": "", “answer": "" },…]})

For each question-answer pair, an additional regular expression is employed to extract the respective figure, equation, or table number following the keywords ‘Figure’, ‘Fig’, ‘Equation’, ‘Eq’, or ‘Table’.

pattern = r’\d+’numbers = re.findall(pattern, input_string)

3.2.5 Step 5 - Crowdworker Cleanup

A significant portion of academic reviews and rebuttals on OpenReview pertain to clarification questions and fixing typos. While the LLM processing removes most of these spurious questions and answers, to further ensure the quality of the dataset we employ crowd workers from Amazon Mechanical Turk to ascertain whether a question-answer pair about an equation, table, or figure is technical in nature or asks to fix a typo. For each question-answer pair, two crowdworkers provide feedback on whether the question is technical in nature or not and only those question-answer pairs with a consensus are retained.

3.2.6 Step 6 - Contextualizing QA Pairs

In addition to extracting question-answer pairs from OpenReview, cPAPERS associates these question-answer pairs with referring text from the LaTeXsource on arXiv.Often, the text surrounding figures, equations, and tables in scientific documents provides additional clarifying information. Regular expressions are employed to locate all instances of figures, equation, and tables in the LaTeXsource using the \begin{figure}, \begin{equation}, and \begin{table} environments.

Equation: The equation is obtained by extracting content enclosed between \begin{equation} and \end{equation} tags. We provide all equations from the LaTeXsource for a question-answer pair.

Table: The table is extracted from the content enclosed between \begin{tabular} and \end{tabular} tags. We provide all tables from the LaTeXsource for a question-answer pair.

Figure: First, the content between \begin{figure} and \end{figure} tags is extracted. From this content, we obtain the caption (using \caption) and the figure path (using \includegraphics).

Additionally, for each modality, we provide context and references:

Context: The context represents the paragraph of text preceding and the paragraph of text following the equation, table, or figure referred to in the question-answer pair.

References: In addition, we report all referring text in a document associated with the equation, table, or figure in the question-answer pair. Within each environment, we search for a label (\label) and if it exists, we locate all references by searching for text using the label in LaTeX’s \ref command.

Reviews of academic pre-prints provide a source for high-quality document-grounded questions and answers and success on the task requires contextual and multimodal understanding. However, one of the main challenges with matching questions and answers from OpenReview with their corresponding multimodal context from arXiv is that while reviewer and author references to equations, tables, and figures are purely ordinal (e.g. Equation 3, Figure 4, Table 2), the LaTeXsource does not provide any reliable approach to associate this ordinal position with the multimodal context in the final compiled document, specifically for equations and tables. Additionally, arXiv contains multiple versions of papers with no guarantees on consistency with the submission to OpenReview. As a result, the cPAPERS dataset reports both the position of the equation, table, or figure referred to in the question, and all equations and tables from that specific paper. Section 4 contains additional details on how this information is used in a baseline approach.

3.2.7 Step 7 - Additional Post-processing of Figures

Authors frequently utilize different graphical formats in their papers. To ensure consistency in our dataset, we convert all .pdf and .eps files to .png format using ImageMagick 444https://imagemagick.org/.

4 Baseline Approaches

We experiment with zero-shot prompting and parameter-efficient fine-tuning for the baseline approaches. Zero-shot prompting involves querying pre-trained LLMs to answer the questions in the cPAPERS dataset without additional fine-tuning. In this approach, we use off-the-shelf Transformer decoder LLMs already pre-trained using a causal language modeling objective followed by supervised fine-tuning and reinforcement learning with human feedback to follow instructions. The LLM is not fine-tuned on cPAPERS and we do not provide examples in-context.

In addition, we experiment with parameter-efficient fine-tuning using QLoRA [14] to better align the model with the style of responses in the cPAPERS dataset. An LLM is fine-tuned independently for each modality (equation, table, or figure).

We serialize equations and tables using the corresponding LaTeXrepresentations from the .tex source on arXiv. As previously indicated in Section 3.2.6, pairing questions and answers from OpenReview with LaTeXsources from arXiv presents a challenge with possible inconsistencies between the references in questions and answers to the corresponding sources in LaTeX. Therefore, we conduct question-answering experiments under two settings - utilizing all multimodal content, or utilizing a smaller subset of neighboring weakly-grounded multimodal content. While using all multimodal content (equations or tables) ensures that the correct equation (or table) referred to by the question-answer pair is provided to the model, limitations on context length constrain the model from accurately using this information. Evidenced by prior work demonstrating the retrieval capabilities of generative LLMs when provided weakly-grounded context [47], we reconcile this situation by providing the model with a subset of equations or tables from the entire paper. First, we assign each equation or table with an ordinal position based on each instance of a \begin{equation} or \begin{table} in the LaTeXsource. Then, supposing the question-answer pair refers to equationi, the LLM is provided a subset of equations: {equationi1,equationi,equationi+1}subscriptequation𝑖1subscriptequation𝑖subscriptequation𝑖1\{\text{equation}_{i-1},\text{equation}_{i},\text{equation}_{i+1}\}{ equation start_POSTSUBSCRIPT italic_i - 1 end_POSTSUBSCRIPT , equation start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT , equation start_POSTSUBSCRIPT italic_i + 1 end_POSTSUBSCRIPT }. We utilize one-sided grounding in the situation where grounding from either boundary does not exist.

5 Results

The following sections contain results of the baseline approaches on the development and test splits. We report results from automatic evaluation - ROUGE [28] 555https://huggingface.co/spaces/evaluate-metric/rouge, METEOR [4] 666https://huggingface.co/spaces/evaluate-metric/meteor, and BERTScore 777https://huggingface.co/spaces/evaluate-metric/bertscore [58].

Zero-shot. First, we experiment with zero-shot language modeling with Llama-2-70B [50] to answer questions on cPAPERS-EQNS, cPAPERS-TBLS, and cPAPERS-FIGS. The results from Tables 2 and 3 indicate that utilizing the neighboring equations/tables provides a significant improvement over using all equations/tables to answer questions in the dataset, evidenced by up to 2x improvement in ROUGE, METEOR, and BERTScores. Additionally, the results from Table2 indicate that utilizing the question alone without neighboring textual context (paragraphs of text surrounding the equation) results in best ROUGE and METEOR scores on cPAPERS-EQNS, whereas utilizing referring text yields the best scores on these metrics in cPAPERS-TBLS when compared to using the question alone to generate answers (Table 3).We posit this is because referring text in the document is predominantly used to summarize tabular results.

ModalitySettingROUGE-1ROUGE-2ROUGE-LMETEORBERTScore
Question (Q)-0.1940.0650.1440.2400.825
Q+EquationNeighboring0.1900.0630.1390.2450.821
All0.1700.0560.1230.2180.740
Q+ContextNeighboring0.1860.0630.1370.2370.809
All0.0790.0270.0580.1020.345
Q+ReferencesNeighboring0.1760.0610.1290.2230.764
All0.1120.0370.0820.1430.498

ModalitySettingROUGE-1ROUGE-2ROUGE-LMETEORBERTScore
Question (Q)-0.1920.0580.1360.2320.832
Q+TableNeighboring0.2060.0610.1450.2370.828
All0.1760.0520.1230.1990.697
Q+ContextNeighboring0.2020.0620.1420.2410.828
All0.1600.0500.1140.1950.666
Q+ReferencesNeighboring0.2070.0640.1440.2430.829
All0.1860.0570.1300.2230.758

Table 4 contains results for zero-shot language modeling on cPAPERS-FIGS. In this situation, utilizing context performs the best across metrics.

ModalityROUGE-1ROUGE-2ROUGE-LMETEORBERTScore
Question (Q)0.1850.0650.1370.2380.833
Q+Caption0.2000.0740.1490.2480.837
Q+Context0.2080.0760.1550.2540.837
Q+References0.2050.0750.1540.2510.837

Fine-tuning. We also report results by fine-tuning Llama-2-7B using QLoRA [14]. Motivated by the zero-shot xperiments, we only conduct fine-tuning using the neighboring setting for cPAPERS-EQNS and -TBLS.

Results of fine-tuning on cPAPERS-EQNS are detailed in Table 5. Utilizing the question with the equation outperforms the question alone or using the question along with referring text or the context on ROUGE-1 and ROUGE-L scores, METEOR, and BERTScore. However, the ANOVA test shows that the difference is not statistically significant.

ModalityROUGE-1ROUGE-2ROUGE-LMETEORBERTScore
Question (Q)0.3090.1380.2480.2150.860
Q + Equation0.3170.1340.2510.2230.861
Q + Context0.2970.1260.2350.2210.817
Q + References0.2830.1220.2240.2170.777

In contrast, the results from finetuning cPAPERS-TBLS and cPAPERS-FIGS in Tables 6 and 7 indicate that using just the question results in the best performance across most metrics. We believe that this is a result of the nature of questions and answers in reviews where authors often utilize tables and figures to supplement reviewer questions. This is supported by Table 8. In the cPAPERS-EQNS split of the dataset, 55.9% of questions refer to an equation while 77.6% of answers refer to an equation. On cPAPERS-TBLS and cPAPERS-FIGS, this difference is exacerbated, with answers referring to a table or a figure more often and questions referring to them less often when compared to equations.

ModalityROUGE-1ROUGE-2ROUGE-LMETEORBERTScore
Question (Q)0.3150.1210.2350.2180.869
Q + Table0.2930.1070.2180.2120.820
Q + Context0.2940.1060.2160.2250.838
Q + References0.2920.1060.2140.2180.816
ModalityROUGE-1ROUGE-2ROUGE-LMETEORBERTScore
Question (Q)0.3290.1550.2690.2500.859
Q + Figure0.3220.1470.2600.2460.868
Q + Context0.3210.1400.2560.2500.858
Q + References0.3110.1310.2440.2470.843
SplitQuestion (%)Answer (%)
cPAPERS-EQNS55.9477.65
cPAPERS-TBLS46.4779.51
cPAPERS-FIGS54.1680.48

6 Conclusion

This paper introduces the cPAPERS dataset, which addresses the shortcomings of prior datasets as a multimodal, conversational dataset grounded in scientific documents. Leveraging reviews of academic papers grounded in equations, figures, and tables, as well as their associated references from scientific documents available on arXiv, this dataset aims to further advance the development of conversational assistants capable of situated and multimodal interactive conversation within scientific papers.We conduct a series of experiments with zero-shot prompting to benchmark state-of-the-art models performance and perform several parameter-efficient fine-tuning for the baseline approaches. The experimental results provide valuable insights and create opportunities for enhancing and innovating the development of future conversational AI assistant.

7 Discussion

This paper introduces an innovative and scalable approach for collecting a large dataset of situated and multimodal interactive conversations related to scientific papers. This approach has the potential to facilitate the collection of extensive question-answer pairs across a wide range of scientific fields with ease. Collecting such a large and diverse dataset would benefit the research community by providing opportunities to develop more advanced AI-based research assistants for scientific research.

There may be concerns that creating an AI assistant to help automate the scientific discovery process could diminish the role of human scientists. While this may be a possible use of this dataset, our intention is to develop an AI research assistant that amplifies human scientists, empowering them to achieve greater scientific discoveries more efficiently.

Limitations. The key limitation of this dataset is the presence of mismatched figures, tables, or equations across different versions of the manuscripts. The .tex files of the paper on arXiv often undergo multiple revisions, and comments on Open Review are typically specific to a particular version. Authors frequently make changes in response to reviewer comments, which may involve additions, removals, or reordering of figures, equations, or tables.This presents a potential mismatch between the figures, equations, or tables and the question-and-answer pairs, thereby introducing additional challenges for language modeling.

Acknowledgments and Disclosure of Funding

This work was supported by NSF IIS-2112633 and by CoCoSys, one of seven centers in JUMP 2.0, a Semiconductor Research Corporation (SRC) program sponsored by DARPA.

References

  • [1]Aishwarya Agrawal, Jiasen Lu, Stanislaw Antol, Margaret Mitchell, C.Lawrence Zitnick, Dhruv Batra, and Devi Parikh.VQA: Visual Question Answering.arXiv:1505.00468 [cs], October 2016.arXiv: 1505.00468.
  • [2]Mubashara Akhtar, Abhilash Shankarampeta, Vivek Gupta, Arpit Patil, Oana Cocarascu, and Elena Simperl.Exploring the numerical reasoning capabilities of language models: A comprehensive analysis on tabular data.In Houda Bouamor, Juan Pino, and Kalika Bali, editors, Findings of the Association for Computational Linguistics: EMNLP 2023, pages 15391–15405, Singapore, December 2023. Association for Computational Linguistics.
  • [3]Rami Aly, Zhijiang Guo, MichaelSejr Schlichtkrull, James Thorne, Andreas Vlachos, Christos Christodoulopoulos, Oana Cocarascu, and Arpit Mittal.The fact extraction and VERification over unstructured and structured information (FEVEROUS) shared task.In Rami Aly, Christos Christodoulopoulos, Oana Cocarascu, Zhijiang Guo, Arpit Mittal, Michael Schlichtkrull, James Thorne, and Andreas Vlachos, editors, Proceedings of the Fourth Workshop on Fact Extraction and VERification (FEVER), pages 1–13, Dominican Republic, November 2021. Association for Computational Linguistics.
  • [4]Satanjeev Banerjee and Alon Lavie.METEOR: An automatic metric for MT evaluation with improved correlation with human judgments.In Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, pages 65–72, Ann Arbor, Michigan, June 2005. Association for Computational Linguistics.
  • [5]Lukas Blecher, Guillem Cucurull, Thomas Scialom, and Robert Stojnic.Nougat: Neural Optical Understanding for Academic Documents, August 2023.arXiv:2308.13418 [cs].
  • [6]Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, JaredD Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, etal.Language models are few-shot learners.Advances in Neural Information Processing Systems, 33:1877–1901, 2020.
  • [7]Wenhu Chen, Hongmin Wang, Jianshu Chen, Yunkai Zhang, Hong Wang, Shiyang Li, Xiyou Zhou, and WilliamYang Wang.Tabfact: A large-scale dataset for table-based fact verification.In International Conference on Learning Representations, 2020.
  • [8]Zhiyu Chen, Wenhu Chen, Charese Smiley, Sameena Shah, Iana Borova, Dylan Langdon, Reema Moussa, Matt Beane, Ting-Hao Huang, Bryan Routledge, and WilliamYang Wang.FinQA: A dataset of numerical reasoning over financial data.In Marie-Francine Moens, Xuanjing Huang, Lucia Specia, and Scott Wen-tau Yih, editors, Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 3697–3711, Online and Punta Cana, Dominican Republic, November 2021. Association for Computational Linguistics.
  • [9]Zhoujun Cheng, Haoyu Dong, Zhiruo Wang, Ran Jia, Jiaqi Guo, Yan Gao, Shi Han, Jian-Guang Lou, and Dongmei Zhang.HiTab: A hierarchical table dataset for question answering and natural language generation.In Smaranda Muresan, Preslav Nakov, and Aline Villavicencio, editors, Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1094–1110, Dublin, Ireland, May 2022. Association for Computational Linguistics.
  • [10]Ting-Rui Chiang and Yun-Nung Chen.Semantically-aligned equation generation for solving and reasoning math word problems.In Jill Burstein, Christy Doran, and Thamar Solorio, editors, Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 2656–2668, Minneapolis, Minnesota, June 2019. Association for Computational Linguistics.
  • [11]Christopher Clark and Santosh Divvala.Extracting Figures, Tables, and Captions from Computer Science Papers.
  • [12]Abhishek Das, Satwik Kottur, Khushi Gupta, Avi Singh, Deshraj Yadav, José M.F. Moura, Devi Parikh, and Dhruv Batra.Visual Dialog.arXiv:1611.08669 [cs], August 2017.arXiv: 1611.08669.
  • [13]Yang Deng, Wenqiang Lei, Wenxuan Zhang, Wai Lam, and Tat-Seng Chua.PACIFIC: Towards proactive conversational question answering over tabular and textual data in finance.In Yoav Goldberg, Zornitsa Kozareva, and Yue Zhang, editors, Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 6970–6984, Abu Dhabi, United Arab Emirates, December 2022. Association for Computational Linguistics.
  • [14]Tim Dettmers, Artidoro Pagnoni, Ari Holtzman, and Luke Zettlemoyer.QLoRA: Efficient Finetuning of Quantized LLMs.In A.Oh, T.Neumann, A.Globerson, K.Saenko, M.Hardt, and S.Levine, editors, Advances in Neural Information Processing Systems, volume36, pages 10088–10115. Curran Associates, Inc., 2023.
  • [15]Julian Eisenschlos, Syrine Krichene, and Thomas Müller.Understanding tables with intermediate pre-training.In Trevor Cohn, Yulan He, and Yang Liu, editors, Findings of the Association for Computational Linguistics: EMNLP 2020, pages 281–296, Online, November 2020. Association for Computational Linguistics.
  • [16]Zhengcong Fei, Zekang Li, Jinchao Zhang, Yang Feng, and Jie Zhou.Towards Expressive Communication with Internet Memes: A New Multimodal Conversation Dataset and Benchmark.arXiv:2109.01839 [cs], September 2021.arXiv: 2109.01839.
  • [17]Heng Gong, Yawei Sun, Xiaocheng Feng, Bing Qin, Wei Bi, Xiaojiang Liu, and Ting Liu.TableGPT: Few-shot table-to-text generation with table structure reconstruction and content matching.In Donia Scott, Nuria Bel, and Chengqing Zong, editors, Proceedings of the 28th International Conference on Computational Linguistics, pages 1978–1988, Barcelona, Spain (Online), December 2020. International Committee on Computational Linguistics.
  • [18]Ming Gong, Xin Wei, Diane Oyen, Jian Wu, Martin Gryder, and Liping Yang.Recognizing Figure Labels in Patents.In SDU@ AAAI, 2021.
  • [19]Zihui Gu, JuFan, Nan Tang, Preslav Nakov, Xiaoman Zhao, and Xiaoyong Du.PASTA: Table-operations aware fact verification via sentence-table cloze pre-training.In Yoav Goldberg, Zornitsa Kozareva, and Yue Zhang, editors, Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 4971–4983, Abu Dhabi, United Arab Emirates, December 2022. Association for Computational Linguistics.
  • [20]Darryl Hannan, Akshay Jain, and Mohit Bansal.ManyModalQA: Modality Disambiguation and QA over Diverse Inputs.arXiv:2001.08034 [cs], January 2020.arXiv: 2001.08034.
  • [21]Darryl Hannan, Akshay Jain, and Mohit Bansal.Manymodalqa: Modality disambiguation and qa over diverse inputs.In Proceedings of the AAAI Conference on Artificial Intelligence, volume34, pages 7879–7886, 2020.
  • [22]Jonathan Herzig, Thomas Müller, Syrine Krichene, and Julian Eisenschlos.Open domain question answering over tables via dense retrieval.In Kristina Toutanova, Anna Rumshisky, Luke Zettlemoyer, Dilek Hakkani-Tur, IzBeltagy, Steven Bethard, Ryan Cotterell, Tanmoy Chakraborty, and Yichao Zhou, editors, Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 512–519, Online, June 2021. Association for Computational Linguistics.
  • [23]Jonathan Herzig, PawelKrzysztof Nowak, Thomas Müller, Francesco Piccinno, and Julian Eisenschlos.TaPas: Weakly supervised table parsing via pre-training.In Dan Jurafsky, Joyce Chai, Natalie Schluter, and Joel Tetreault, editors, Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 4320–4333, Online, July 2020. Association for Computational Linguistics.
  • [24]Ting-Yao Hsu, CLee Giles, and Ting-Hao Huang.SciCap: Generating captions for scientific figures.In Marie-Francine Moens, Xuanjing Huang, Lucia Specia, and Scott Wen-tau Yih, editors, Findings of the Association for Computational Linguistics: EMNLP 2021, pages 3258–3264, Punta Cana, Dominican Republic, November 2021. Association for Computational Linguistics.
  • [25]Jinhao Jiang, Kun Zhou, Zican Dong, Keming Ye, Xin Zhao, and Ji-Rong Wen.StructGPT: A general framework for large language model to reason over structured data.In Houda Bouamor, Juan Pino, and Kalika Bali, editors, Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 9237–9251, Singapore, December 2023. Association for Computational Linguistics.
  • [26]Satwik Kottur and Seungwhan Moon.Overview of situated and interactive multimodal conversations (SIMMC) 2.1 track at DSTC 11.In Yun-Nung Chen, Paul Crook, Michel Galley, Sarik Ghazarian, Chulaka Gunasekara, Raghav Gupta, Behnam Hedayatnia, Satwik Kottur, Seungwhan Moon, and Chen Zhang, editors, Proceedings of The Eleventh Dialog System Technology Challenge, pages 235–241, Prague, Czech Republic, September 2023. Association for Computational Linguistics.
  • [27]Esther Landhuis.Scientific literature: Information overload.Nature, 535(7612):457–458, 2016.
  • [28]Chin-Yew Lin.ROUGE: A Package for Automatic Evaluation of Summaries.In Text Summarization Branches Out, pages 74–81, Barcelona, Spain, July 2004. Association for Computational Linguistics.
  • [29]Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and CLawrence Zitnick.Microsoft coco: Common objects in context.In Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, pages 740–755. Springer, 2014.
  • [30]Qian Liu, Bei Chen, Jiaqi Guo, Morteza Ziyadi, Zeqi Lin, Weizhu Chen, and Jian-Guang Lou.Tapex: Table pre-training via learning a neural sql executor.arXiv preprint arXiv:2107.07653, 2021.
  • [31]Kenneth Marino, Mohammad Rastegari, Ali Farhadi, and Roozbeh Mottaghi.OK-VQA: A Visual Question Answering Benchmark Requiring External Knowledge.In Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
  • [32]Seungwhan Moon, Satwik Kottur, PaulA. Crook, Ankita De, Shivani Poddar, Theodore Levin, David Whitney, Daniel Difranco, Ahmad Beirami, Eunjoon Cho, Rajen Subba, and Alborz Geramifard.Situated and interactive multimodal conversations.CoRR, abs/2006.01460, 2020.
  • [33]Nasrin Mostafazadeh, Chris Brockett, Bill Dolan, Michel Galley, Jianfeng Gao, GeorgiosP. Spithourakis, and Lucy Vanderwende.Image-Grounded Conversations: Multimodal Context for Natural Question and Response Generation.arXiv:1701.08251 [cs], April 2017.arXiv: 1701.08251.
  • [34]Kai Nakamura, Sharon Levy, Yi-Lin Tuan, Wenhu Chen, and WilliamYang Wang.HybriDialogue: An information-seeking dialogue dataset grounded on tabular and textual data.In Smaranda Muresan, Preslav Nakov, and Aline Villavicencio, editors, Findings of the Association for Computational Linguistics: ACL 2022, pages 481–492, Dublin, Ireland, May 2022. Association for Computational Linguistics.
  • [35]Linyong Nan, Chiachun Hsieh, Ziming Mao, XiVictoria Lin, Neha Verma, Rui Zhang, Wojciech Kryściński, Hailey Schoelkopf, Riley Kong, Xiangru Tang, Mutethia Mutuma, Ben Rosand, Isabel Trindade, Renusree Bandaru, Jacob Cunningham, Caiming Xiong, Dragomir Radev, and Dragomir Radev.FeTaQA: Free-form table question answering.Transactions of the Association for Computational Linguistics, 10:35–49, 2022.
  • [36]Panupong Pasupat and Percy Liang.Compositional semantic parsing on semi-structured tables.In Chengqing Zong and Michael Strube, editors, Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 1470–1480, Beijing, China, July 2015. Association for Computational Linguistics.
  • [37]Shuai Peng, KeYuan, Liangcai Gao, and Zhi Tang.Mathbert: A pre-trained model for mathematical formula understanding.arXiv preprint arXiv:2105.00377, 2021.
  • [38]Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, etal.Language models are unsupervised multitask learners.OpenAI blog, 1(8):9, 2019.
  • [39]Krithika Ramesh, Arnav Chavan, Shrey Pandit, and Sunayana Sitaram.A comparative study on the impact of model compression techniques on fairness in language models.In Anna Rogers, Jordan Boyd-Graber, and Naoaki Okazaki, editors, Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 15762–15782, Toronto, Canada, July 2023. Association for Computational Linguistics.
  • [40]BenjaminZ Reichman, Anirudh Sundar, Christopher Richardson, Tamara Zubatiy, Prithwijit Chowdhury, Aaryan Shah, Jack Truxal, Micah Grimes, Dristi Shah, WooJu Chee, and others.Outside knowledge visual question answering version 2.0.In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1–5. IEEE, 2023.
  • [41]Soumajyoti Sarkar and Leonard Lausen.Testing the limits of unified sequence to sequence llm pretraining on diverse table data tasks, 2023.
  • [42]Alexander Scarlatos and Andrew Lan.Tree-based representation and generation of natural and mathematical language.In Anna Rogers, Jordan Boyd-Graber, and Naoaki Okazaki, editors, Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 3714–3730, Toronto, Canada, July 2023. Association for Computational Linguistics.
  • [43]Kurt Shuster, Samuel Humeau, Antoine Bordes, and Jason Weston.Image Chat: Engaging Grounded Conversations.arXiv:1811.00945 [cs], April 2020.arXiv: 1811.00945.
  • [44]Yuan Sui, Mengyu Zhou, Mingjie Zhou, Shi Han, and Dongmei Zhang.Table meets llm: Can large language models understand structured table data? a benchmark and empirical study.In Proceedings of the 17th ACM International Conference on Web Search and Data Mining, WSDM ’24, page 645–654, New York, NY, USA, 2024. Association for Computing Machinery.
  • [45]Anirudh Sundar and Larry Heck.Multimodal conversational AI: A survey of datasets and approaches.In Bing Liu, Alexandros Papangelis, Stefan Ultes, Abhinav Rastogi, Yun-Nung Chen, Georgios Spithourakis, Elnaz Nouri, and Weiyan Shi, editors, Proceedings of the 4th Workshop on NLP for Conversational AI, pages 131–147, Dublin, Ireland, May 2022. Association for Computational Linguistics.
  • [46]Anirudh Sundar, Christopher Richardson, William Gay, and Larry Heck.itbls: A dataset of interactive conversations over tabular information.arXiv preprint arXiv:2404.12580, 2024.
  • [47]AnirudhS. Sundar and Larry Heck.cTBLS: Augmenting large language models with conversational tables.In Yun-Nung Chen and Abhinav Rastogi, editors, Proceedings of the 5th Workshop on NLP for Conversational AI (NLP4ConvAI 2023), pages 59–70, Toronto, Canada, July 2023. Association for Computational Linguistics.
  • [48]Hao Tan, Chen-Tse Tsai, Yujie He, and Mohit Bansal.Scientific Chart Summarization: Datasets and Improved Text Modeling.In SDU@AAAI, 2022.
  • [49]Bart Thomee, DavidA. Shamma, Gerald Friedland, Benjamin Elizalde, Karl Ni, Douglas Poland, Damian Borth, and Li-Jia Li.YFCC100M: The New Data in Multimedia Research.Communications of the ACM, 59(2):64–73, January 2016.arXiv: 1503.01817.
  • [50]Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, etal.Llama: Open and efficient foundation language models.arXiv preprint arXiv:2302.13971, 2023.
  • [51]Zichao Wang, Andrew Lan, and Richard Baraniuk.Math word problem generation with mathematical consistency and problem context constraints.In Marie-Francine Moens, Xuanjing Huang, Lucia Specia, and Scott Wen-tau Yih, editors, Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 5986–5999, Online and Punta Cana, Dominican Republic, November 2021. Association for Computational Linguistics.
  • [52]Zichao Wang, Mengxue Zhang, RichardG Baraniuk, and AndrewS Lan.Scientific formula retrieval via tree embeddings.In 2021 IEEE International Conference on Big Data (Big Data), pages 1493–1503. IEEE, 2021.
  • [53]Jingfeng Yang, Aditya Gupta, Shyam Upadhyay, Luheng He, Rahul Goel, and Shachi Paul.TableFormer: Robust transformer modeling for table-text encoding.In Smaranda Muresan, Preslav Nakov, and Aline Villavicencio, editors, Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 528–537, Dublin, Ireland, May 2022. Association for Computational Linguistics.
  • [54]Michihiro Yasunaga and JohnD Lafferty.Topiceq: A joint topic and mathematical equation model for scientific texts.In Proceedings of the AAAI conference on artificial intelligence, volume33, pages 7394–7401, 2019.
  • [55]Pengcheng Yin, Graham Neubig, Wen-tau Yih, and Sebastian Riedel.TaBERT: Pretraining for joint understanding of textual and tabular data.In Dan Jurafsky, Joyce Chai, Natalie Schluter, and Joel Tetreault, editors, Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 8413–8426, Online, July 2020. Association for Computational Linguistics.
  • [56]Liangyu Zha, Junlin Zhou, Liyao Li, Rui Wang, Qingyi Huang, Saisai Yang, Jing Yuan, Changbao Su, Xiang Li, Aofeng Su, etal.Tablegpt: Towards unifying tables, nature language and commands into one gpt.arXiv preprint arXiv:2307.08674, 2023.
  • [57]Tianshu Zhang, Xiang Yue, Yifei Li, and Huan Sun.Tablellama: Towards open large generalist models for tables.arXiv preprint arXiv:2311.09206, 2023.
  • [58]Tianyi Zhang, Varsha Kishore*, Felix Wu*, KilianQ. Weinberger, and Yoav Artzi.BERTScore: Evaluating Text Generation with BERT.In International Conference on Learning Representations, 2020.
  • [59]Yilun Zhao, Yunxiang Li, Chenying Li, and Rui Zhang.MultiHiertt: Numerical reasoning over multi hierarchical tabular and textual data.In Smaranda Muresan, Preslav Nakov, and Aline Villavicencio, editors, Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 6588–6600, Dublin, Ireland, May 2022. Association for Computational Linguistics.
  • [60]Fengbin Zhu, Wenqiang Lei, Youcheng Huang, Chao Wang, Shuo Zhang, Jiancheng Lv, Fuli Feng, and Tat-Seng Chua.TAT-QA: A question answering benchmark on a hybrid of tabular and textual content in finance.In Chengqing Zong, Fei Xia, Wenjie Li, and Roberto Navigli, editors, Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 3277–3287, Online, August 2021. Association for Computational Linguistics.
  • [61]Yuke Zhu, Oliver Groth, Michael Bernstein, and LiFei-Fei.Visual7W: Grounded Question Answering in Images.arXiv:1511.03416 [cs], April 2016.arXiv: 1511.03416.
cPAPERS: A Dataset of Situated and Multimodal Interactive Conversations in Scientific Papers (2024)

References

Top Articles
Latest Posts
Article information

Author: Duncan Muller

Last Updated:

Views: 6093

Rating: 4.9 / 5 (79 voted)

Reviews: 94% of readers found this page helpful

Author information

Name: Duncan Muller

Birthday: 1997-01-13

Address: Apt. 505 914 Phillip Crossroad, O'Konborough, NV 62411

Phone: +8555305800947

Job: Construction Agent

Hobby: Shopping, Table tennis, Snowboarding, Rafting, Motor sports, Homebrewing, Taxidermy

Introduction: My name is Duncan Muller, I am a enchanting, good, gentle, modern, tasty, nice, elegant person who loves writing and wants to share my knowledge and understanding with you.