3260 papers • 126 benchmarks • 313 datasets
When given a query, the goal of this task is to retrieve a relevant table from a (potentially large) collection of tables. The query could be a single sentence (such as a question), or it could also be a conversation. As for the retrieval, the tables could be in the raw form (i.e. the values of each cells), the metadata (such as the title, description), or summary statistics.
(Image credit: Papersgraph)
These leaderboards are used to track progress in table-retrieval-18
Use these libraries to find table-retrieval-18 models and implementations
No subtasks available.
This work introduces and addresses the problem of ad hoc table retrieval: answering a keyword query with a ranked list of tables, and introduces various similarity measures for matching those semantic representations.
Evaluation results show that table embeddings can significantly improve upon the performance of state-of-the-art baselines and be utilized in table-related tasks, row population, column population, and table retrieval.
This paper investigates how to encode table content considering the table structure and input length limit of BERT and proposes an approach that incorporates features from prior literature on table retrieval and jointly trains them with BERT.
Graph-based Table Retrieval (GTR), a generalizable NLTR framework with multi-granular graph representation learning, which incorporates a self-supervised pre-training task based on graph-context matching to enhance the robustness and generalizability of the model.
This work describes the development, characteristics and availability of a test collection for the task of Web table retrieval, which uses a large-scale Web Table Corpora extracted from the Common Crawl and shows that proper usage of context labels can benefit previous table retrieval methods.
This work proposes a semantic table retrieval framework for matching information needs (keyword or table queries) against tables in multiple semantic spaces and introduces various similarity measures for matching those semantic representations.
StruBERT is proposed, a structure-aware BERT model that fuses the textual and structural information of a data table to produce context-aware representations for both textual and tabular content of aData table.
The task of table retrieval is focused on, and it is found that DPR performs well without any table-specific design and training, and even achieves superior results compared to DTR when fine-tuned on properly linearized tables.
Conversational Tables cTBLS is introduced, a three-step architecture to retrieve and generate dialogue responses grounded on retrieved tabular information to improve accuracy and performance and judge informativeness to be 4x better than the previous state-of-the-art.
Adding a benchmark result helps the community track progress.