SkillAgentSearch skills...

Tapas

End-to-end neural table-text understanding models.

Install / Use

/learn @google-research/Tapas
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

TAble PArSing (TAPAS)

Code and checkpoints for training the transformer-based Table QA models introduced in the paper TAPAS: Weakly Supervised Table Parsing via Pre-training.

News

2021/09/15

2021/08/24

  • Added a colab to try predictions on open domain question answering.

2021/08/20

2021/07/23

2021/05/13

2021/03/23

2020/12/17

2020/10/19

  • Small change to WTQ training example creation
    • Questions with ambiguous cell matches will now be discarded
    • This improves denotation accuracy by ~1 point
    • For more details see this issue.
  • Added option to filter table columns by textual overlap with question

2020/10/09

2020/08/26

  • Added a colab to try predictions on WTQ

2020/08/05

  • New pre-trained models (see Data section below)
  • reset_position_index_per_cell: New option that allows to train models that instead of using absolute position indices reset the position index when a new cell starts.

2020/06/10

  • Bump TensorFlow to v2.2

2020/06/08

2020/05/07

  • Added a colab to try predictions on SQA

Installation

The easiest way to try out TAPAS with free GPU/TPU is in our Colab, which shows how to do predictions on SQA.

The repository uses protocol buffers, and requires the protoc compiler to run. You can download the latest binary for your OS here. On Ubuntu/Debian, it can be installed with:

sudo apt-get install protobuf-compiler

Afterwards, clone and install the git repository:

git clone https://github.com/google-research/tapas
cd tapas
pip install -e .

To run the test suite we use the tox library which can be run by calling:

pip install tox
tox

Models

We provide pre-trained models for different model sizes.

The metrics are computed by our tool and not the official metrics of the respective tasks. We provide them so one can verify whether one's own runs are in the right ballpark. They are medians over three individual runs.

Models with intermediate pre-training (2020/10/07).

New models based on the ideas discussed in Understanding tables with intermediate pre-training. Learn more about the methods use here.

WTQ

Trained from Mask LM, intermediate data, SQA, WikiSQL.

Size | Reset | Dev Accuracy | Link -------- | --------| -------- | ---- LARGE | noreset | 0.5062 | tapas_wtq_wikisql_sqa_inter_masklm_large.zip LARGE | reset | 0.5097 | tapas_wtq_wikisql_sqa_inter_masklm_large_reset.zip BASE | noreset | 0.4525 | tapas_wtq_wikisql_sqa_inter_masklm_base.zip BASE | reset | 0.4638 | tapas_wtq_wikisql_sqa_inter_masklm_base_reset.zip MEDIUM | noreset | 0.4324 | tapas_wtq_wikisql_sqa_inter_masklm_medium.zip MEDIUM | reset | 0.4324 | tapas_wtq_wikisql_sqa_inter_masklm_medium_reset.zip SMALL | noreset | 0.3681 | tapas_wtq_wikisql_sqa_inter_masklm_small.zip SMALL | reset | 0.3762 | tapas_wtq_wikisql_sqa_inter_masklm_small_reset.zip MINI | noreset | 0.2783 | tapas_wtq_wikisql_sqa_inter_masklm_mini.zip MINI | reset | 0.2854 | tapas_wtq_wikisql_sqa_inter_masklm_mini_reset.zip TINY | noreset | 0.0823 | tapas_wtq_wikisql_sqa_inter_masklm_tiny.zip TINY | reset | 0.1039 | tapas_wtq_wikisql_sqa_inter_masklm_tiny_reset.zip

WIKISQL

Trained from Mask LM, intermediate data, SQA.

Size | Reset | Dev Accuracy | Link -------- | --------| -------- | ---- LARGE | noreset | 0.8948 | tapas_wikisql_sqa_inter_masklm_large.zip LARGE | reset | 0.8979 | tapas_wikisql_sqa_inter_masklm_large_reset.zip BASE | noreset | 0.8859 | tapas_wikisql_sqa_inter_masklm_base.zip BASE | reset | 0.8855 | tapas_wikisql_sqa_inter_masklm_base_reset.zip MEDIUM | noreset | 0.8766 | tapas_wikisql_sqa_inter_masklm_medium.zip MEDIUM | reset | 0.8773 | tapas_wikisql_sqa_inter_masklm_medium_reset.zip SMALL | noreset | 0.8552 | tapas_wikisql_sqa_inter_masklm_small.zip SMALL | reset | 0.8615 | tapas_wikisql_sqa_inter_masklm_small_reset.zip MINI | noreset | 0.8063 | tapas_wikisql_sqa_inter_masklm_mini.zip MINI | reset | 0.82 | tapas_wikisql_sqa_inter_masklm_mini_reset.zip TINY | noreset | 0.3198 | tapas_wikisql_sqa_inter_masklm_tiny.zip TINY | reset | 0.6046 | tapas_wikisql_sqa_inter_masklm_tiny_reset.zip

TABFACT

Trained from Mask LM, intermediate data.

Size | Reset | Dev Accuracy | Link -------- | --------| -------- | ---- LARGE | noreset | 0.8101 | tapas_tabfact_inter_masklm_large.zip LARGE | reset | 0.8159 | tapas_tabfact_inter_masklm_large_reset.zip BASE | noreset | 0.7856 | [tapas_tabfact_inter_masklm_base.zip](https:

View on GitHub
GitHub Stars1.2k
CategoryEducation
Updated12d ago
Forks215

Languages

Python

Security Score

100/100

Audited on Mar 17, 2026

No findings