ICLR 2022 Paper, SOTA Table Pre-training Model, TAPEX: Table Pre-training via Learning a Neural SQL Executor
The largest sythetic corpus (500M pairs of SQL-Table-Answers) used by TAPEX during pre-training.