Tabular Representation, Noisy Operators, and Impacts on Table Structure Understanding Tasks in LLMs

Table Representation Learning Workshop at NeurIPS 2023 |

Publication | Related File

Large language models (LLMs) are increasingly applied for tabular tasks using in-context learning. The prompt representation for a table may play a role in the LLMs ability to process the table. Inspired by prior work, we generate a collection of self-supervised table structure understanding tasks (e.g. navigate to a cell and row; transpose the table) and evaluate the performance differences when using eight formats. In contrast to past work, we introduce eight noise operations inspired by real-world messy data and adversarial inputs, and show that these can impact LLM performance across formats for different structural understanding tasks.