This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

TransTab: Learning Transferable Tabular Transformers Across Tables

Zifeng Wang
UIUC
zifengw2@illinois.edu
&Jimeng Sun
UIUC
jimeng@illinois.edu
Abstract

Tabular data (or tables) are the most widely used data format in machine learning (ML). However, ML models often assume the table structure keeps fixed in training and testing. Before ML modeling, heavy data cleaning is required to merge disparate tables with different columns. This preprocessing often incurs significant data waste (e.g., removing unmatched columns and samples). How to learn ML models from multiple tables with partially overlapping columns? How to incrementally update ML models as more columns become available over time? Can we leverage model pretraining on multiple distinct tables? How to train an ML model which can predict on an unseen table?

To answer all those questions, we propose to relax fixed table structures by introducing a Transferable Tabular Transformer (TransTab) for tables. The goal of TransTab is to convert each sample (a row in the table) to a generalizable embedding vector, and then apply stacked transformers for feature encoding. One methodology insight is combining column description and table cells as the raw input to a gated transformer model. The other insight is to introduce supervised and self-supervised pretraining to improve model performance. We compare TransTab with multiple baseline methods on diverse benchmark datasets and five oncology clinical trial datasets. Overall, TransTab ranks 1.00, 1.00, 1.78 out of 12 methods in supervised learning, feature incremental learning, and transfer learning scenarios, respectively; and the proposed pretraining leads to 2.3% AUC lift on average over the supervised learning.

1 Introduction

Tabular data are ubiquitous in healthcare, engineering, advertising, and finance [1, 2, 3, 4]. They are often stored in a relational database as tables or spreadsheets. Table rows represent the data samples, and columns represent the feature variables of diverse data types (e.g., categorical, numerical, binary, and textual). Recent works enhance tabular ML modeling using deep networks [5, 6, 7, 8] or designing self-supervision [2, 9, 10, 11]. Those existing works require the same table structure in training and testing data. However, there can be multiple tables sharing partially overlapped columns in the real world. Hence, learning across tables is inapplicable. The traditional remedy is to perform data cleaning by removing non-overlapping columns and mismatched samples before training any ML models, which waste data resources [12, 13, 14]. Therefore, learning across tables with disparate columns and transferring knowledge across tables are crucial to extending the success of deep learning/pretraining to the tabular domain.

Tables are highly structured yet flexible. The first step to achieve learning across tables is to rethink the basic elements in tabular data modeling. In computer vision, the basic elements are pixels [15] or patches, [16, 17]; in natural language processing (NLP), the basic elements are words [18] or tokens [19, 20]. In the tabular domain, it is natural to treat cells in each column as independent elements. Columns are mapped to unique indexes then models take the cell values for training and inference. The premise of this modeling formulation is to keep the same column structure in all the tables. But tables often have divergent protocols where the nomenclatures of columns and cells differ. By contrast, our proposed work contextualizes the columns and cells. For example, previous methods represent a cell valued man under the column gender by 0 referring to the codebook {man:0,woman:1}\{\text{man}:0,\text{woman}:1\}. Our model converts the tabular input into a sequence input (e.g., gender is man), which can be modeled with downstream sequence models. We argue such featurizing protocol is generalizable across tables, thus enabling models to apply to different tables.

In a nutshell, we propose Transferable Transformers for Tabular analysis (TransTab), a versatile tabular learning framework 111Our package is available at https://github.com/RyanWangZf/transtab with documentation at https://transtab.readthedocs.io/en/latest/.. TransTab applies to multiple use cases as shown in Fig. 1. The key contributions behind TransTab are

  • A systematic featurizing pipeline considering both column and cell semantics which is shared as the fundamental protocol across tables.

  • Vertical-Partition Contrastive Learning (VPCL) that enables pretraining on multiple tables and also allows finetuning on target datasets.

As shown by Fig. 1, due to the fixed-column assumption, all existing works only handle supervised learning or pretraining on the same-structure tables. On the contrary, TransTab relaxes this assumption and applies to four additional scenarios, which we will elaborate on in §2.1.

Refer to caption
Figure 1: The demonstration of ML modeling on different tabular data settings. Previous tabular methods only do vanilla supervised training or pretraining on the same table due to they only accept fixed-column tables. By contrast, TransTab covers more new tasks (1) to (4) as it accepts variable-column tables. Details are presented in §2.1.

2 Method

In this section, we present the details of TransTab. Fig. 2 illustrates its workflow including the following key components: 1) The input processor featurizes and embeds arbitrary tabular inputs to token-level embeddings; 2) The stacked gated transformer layers further encode the token-level embeddings; 3) Finally, the learning module includes a classifier trained on labeled data and a projector for contrastive learning. Next we will present the details of each component.

Refer to caption
Figure 2: The demonstration of TransTab framework. the input processor encodes the sample into the token-level embedding 𝐄{\mathbf{E}}; the [cls] embedding 𝐳[cls]\mathbf{z}^{[cls]} in the representation 𝐙L{\mathbf{Z}}^{L} after LL gated transformer layers is used for prediction and learning. In supervised learning, 𝐳[cls]\mathbf{z}^{[cls]} is leveraged by a classifier to make predictions of target yy; in contrastive learning, the projected 𝐳^[cls]\hat{\mathbf{z}}^{[cls]} is is used for self or supervised contrastive loss.

2.1 Application scenarios of TransTab

Before presenting our method in details, we first introduce four novel applications scenarios which are tractable by TransTab, as shown in Fig. 1. Suppose we aim to predict the treatment efficacy for breast cancer trials using multiple clinical trial tables, here are several scenarios we often encounter.

S(1) Transfer learning. We collect data tables from multiple cancer trials for testing the efficacy of the same drug on different patients. These tables were designed independently with overlapping columns. How do we learn ML models for one trial by leveraging tables from all trials?

S(2) Incremental learning. Additional columns might be added over time. For example, additional features are collected across different trial phases. How do we update the ML models using tables from all trial phases?

S(3) Pretraining+Finetuning. The trial outcome label (e.g., mortality) might not be always available from all table sources. Can we benefit pretraining on those tables without labels? How do we finetune the model on the target table with labels?

S(4) Zero-shot inference. We model the drug efficacy based on our trial records. The next step is to conduct inference with the model to find patients that can benefit from the drug. However, patient tables do not share the same columns as trial tables so direct inference is not possible.

Overall, we witness that the assumption of fixed table structure is the obstacle to use ML for various applications. Next we will present TransTab and demonstrate how it addresses these scenarios.

2.2 Input processor for columns and cells

We build the input processor (1) to accept variable-column tables (2) to retain knowledge across tabular datasets. The idea is to convert tabular data (cells in columns) into a sequence of semantically encoded tokens. We utilize the following observation to create the sequence: the column description (e.g., column name) decides the meaning of cells in that column. For example, if a cell in column smoking history has value 11, it indicates the individual has a smoking history. Similarly, cell value 6060 in column weight indicates 60 kg in weight instead of 60 years old. Motivated by the discussion, we propose to include column names into the tabular modeling. As a result, TransTab treats any tabular data as the composition of three elements: text (for categorical & textual cells and column names), continuous values (for numerical cells), and boolean values (for binary cells) . Fig. 2 illustrates a visual example of how these elements are leveraged to process the four basic types of features: categorical/textual cat, binary bin, and numerical num.

Categorical/Textual feature. A category or textual feature contains a sequence of text tokens. For the categorical feature cat, we concatenate the column name with the feature value xcx_{c}, which forms as a sequence of tokens. This sentence is then tokenized and matched to the token embedding matrix to generate the feature embedding 𝐄cnc×d{\mathbf{E}}_{c}\in\mathbb{R}^{n_{c}\times d} where dd is the embedding dimension and ncn_{c} is the number of tokens.

Binary feature. The binary feature bin is usually an assertive description and its value xb{0,1}x_{b}\in\{0,1\}. If xb=1x_{b}=1, then bin is tokenized and encoded to the embeddings 𝐄bnb×d{\mathbf{E}}_{b}\in\mathbb{R}^{n_{b}\times d}; if not, it will not be processed to the subsequent steps. This design significantly reduces the computational and memory cost when the inputs have high-dimensional and sparse one-hot features.

Numerical feature. We do not concatenate column names and values for numerical feature because the tokenization-embedding paradigm was notoriously known to be bad at discriminating numbers [21]. Instead, we process them separately. num is encoded as same as cat and bin to get 𝐄u,colnu×d{\mathbf{E}}_{u,col}\in\mathbb{R}^{n_{u}\times d}. We then multiply the numerical features with the column embedding to yield the numerical embedding as 𝐄u=xu×𝐄u,col{\mathbf{E}}_{u}=x_{u}\times{\mathbf{E}}_{u,col}222xux_{u} is standardized or normalized in preprocessing., which we identify gets an edge on more complicated numerical embedding techniques empirically.

At last, 𝐄c{\mathbf{E}}_{c}, 𝐄u{\mathbf{E}}_{u}, 𝐄b{\mathbf{E}}_{b} all pass the layer normalization [22] and the same linear layer to be aligned to the same space, then are concatenated with [cls] embedding to yield 𝐄=𝐄~c𝐄~u𝐄~b𝐞[cls]{\mathbf{E}}=\tilde{{\mathbf{E}}}_{c}\otimes\tilde{{\mathbf{E}}}_{u}\otimes\tilde{{\mathbf{E}}}_{b}\otimes\mathbf{e}^{[cls]}.

As a result, all cell values are contextualized regarding the corresponding column properties thus the semantic meaning of one element can vary depending on the context composition. This formulation benefits the knowledge transfer across tables a lot. For example, previously smoked depicts the same thing as smoking history. Previous methods never capture this connection while it is possible for TransTab to learn to recognize that 11 under both columns are equivalent.

2.3 Gated transformers

The gated tabular transformer is an adaption of the classical transformer in NLP [23]. It consists of two main components: multi-head self-attention layer and gated feedforward layers. The input representation 𝐙l{\mathbf{Z}}^{l} at the ll-th layer is first adopted for exploring interactions between features:

𝐙attl=MultiHeadAttn(𝐙l)=[head1,head2,,headh]𝐖O,\displaystyle{\mathbf{Z}}^{l}_{\text{att}}=\texttt{MultiHeadAttn}({\mathbf{Z}}^{l})=[\texttt{head}_{1},\texttt{head}_{2},\dots,\texttt{head}_{h}]{\rm\mathbf{W}}^{O}, (1)
headi=Attention(𝐙l𝐖iQ,𝐙l𝐖iK,𝐙l𝐖iV),\displaystyle\texttt{head}_{i}=\texttt{Attention}({\mathbf{Z}}^{l}{\rm\mathbf{W}}^{Q}_{i},{\mathbf{Z}}^{l}{\rm\mathbf{W}}^{K}_{i},{\mathbf{Z}}^{l}{\rm\mathbf{W}}^{V}_{i}), (2)

where 𝐙0=𝐄{\mathbf{Z}}^{0}={\mathbf{E}} at the first layer; 𝐖Od×d{\rm\mathbf{W}}^{O}\in\mathbb{R}^{d\times d}; {𝐖iQ,𝐖iK,𝐖iV}\{{\rm\mathbf{W}}^{Q}_{i},{\rm\mathbf{W}}^{K}_{i},{\rm\mathbf{W}}^{V}_{i}\} are weight matrices (in d×dh\mathbb{R}^{d\times\frac{d}{h}}) of query, key, value of the ii-th head self-attention module.

The multi-head attention output 𝐙attl{\mathbf{Z}}^{l}_{\text{att}} is further transformed by a token-wise gating layer as 𝐠l=σ(𝐙attl𝐰G)\mathbf{g}^{l}=\sigma({\mathbf{Z}}_{\text{att}}^{l}\mathbf{w}^{G}), where σ()\sigma(\cdot) is a sigmoid function; 𝐠l[0,1]n\mathbf{g}^{l}\in[0,1]^{n} controls the magnitude of each token embedding before 𝐙att{\mathbf{Z}}_{\text{att}} goes to the linear projection. This gates then filters the linear layer output

𝐙l+1=Linear((𝐠l𝐙attl)Linear(𝐙attl)){\mathbf{Z}}^{l+1}=\texttt{Linear}\left((\mathbf{g}^{l}\odot{\mathbf{Z}}^{l}_{\text{att}})\oplus\texttt{Linear}({\mathbf{Z}}_{\text{att}}^{l})\right) (3)

to obtain the transformer output 𝐙l+1n×d{\mathbf{Z}}^{l+1}\in\mathbb{R}^{n\times d}. This mechanism is learnt to focus on important features by redistributing the attention on tokens. The final [cls] embedding 𝐳[cls]\mathbf{z}^{[cls]} at the LL-th layer is used by the classifier for prediction.

2.4 Self-supervised and supervised pretraining of TransTab

The input processor accepts variable-column tables, which opens the door for tabular pretraining on heterogeneous tables. In detail, TransTab is feasible for self-supervised and supervised pretraining.

Self-supervised VPCL. Most SSL tabular methods work on the whole fixed set of columns [2, 24, 11], which take high computational costs and are prone to overfitting. Instead, we take tabular vertical partitions to build positive and negative samples for CL under the hypothesis that the powerful representation should model view-invariant factors. In detail, we subset columns as illustrated by Fig. 3 where Self-VPCL is on the top right. Suppose a sample 𝐱i={𝐯i1,,𝐯iK}\mathbf{x}_{i}=\{\mathbf{v}_{i}^{1},\dots,\mathbf{v}_{i}^{K}\} with KK partitions 𝐯ik\mathbf{v}_{i}^{k}. Neighbouring partitions can have overlapping regions which are justified by the percentage of columns of the partition. Self-VPCL takes partitions from the same sample as the positive and others as the negative:

(𝐗)=i=1Bk=1KkkKlogexpψ(𝐯ik,𝐯ik)j=1Bk=1Kexpψ(𝐯ik,𝐯jk),\ell({\rm\mathbf{X}})=-\sum_{i=1}^{B}\sum_{k=1}^{K}\sum_{k^{\prime}\neq k}^{K}\log\frac{\exp\psi(\mathbf{v}_{i}^{k},\mathbf{v}_{i}^{k^{\prime}})}{\sum_{j=1}^{B}\sum_{k^{\dagger}=1}^{K}\exp\psi(\mathbf{v}_{i}^{k},\mathbf{v}_{j}^{k^{\dagger}})}, (4)

where BB is the batch size; ψ(,)\psi(\cdot,\cdot) is the cosine similarity function. ψ\psi applies to 𝐳^[cls]\hat{\mathbf{z}}^{[cls]} which is the linear projection of partition 𝐯\mathbf{v}’s embedding 𝐳[cls]\mathbf{z}^{[cls]}. Compared with vanilla CL like SCARF [11], Self-VPCL significantly expand the positive and negative sampling for learning more robust and rich embeddings. What is more, this vertical partition sampling is extremely friendly to column-oriented databases [25] which support the fast querying a subset of columns from giant data warehouses. For the sake of computational efficiency, when K>2K>2, we randomly sample two partitions.

Refer to caption
Figure 3: The demonstration of contrastive learning methods (different pieces can either be distinct or be overlapped partially). Self-VPCL: Positive pairs are partitions of the same sample; VPCL: Positive pairs are partitions of the sample belonging to the same class.

Supervised VPCL. When we own labeled tabular data for pretraining, one natural idea would be taking task-specific predicting heads for pretraining on vanilla supervised loss, e.g., cross-entropy loss. In finetuning, these heads are dropped and a new head will be added on top of the pretrained encoder. However, we argue it is suboptimal and may undermine the model transferability. The reason behind is that tabular datasets vary dramatically in size, task definition, and class distributions. Pretraining TransTab using supervised loss inevitably causes the encoder biased to the major tasks and classes. Moreover, the suitable hyperparameter range is often distinct across tabular data when applying supervised loss. The same set of hyperparameters can cause overfitting on one dataset and underfitting on another. Therefore, it is tricky to pick appropriate hyperparameters for pretraining based on vanilla supervised loss.

In this paper, we propose VPCL for pretraining inspired by supervised CL [26] which was proved robust to noise and hyperparameters. As illustrated by Fig. 3, we build positive pairs considering views from the same class except for only from the same sample:

(𝐗,𝐲)=i=1Bj=1Bk=1Kk=1K𝟏{yj=yi}logexpψ(𝐯ik,𝐯jk)j=1Bk=1K𝟏{yjyi}expψ(𝐯ik,𝐯jk).\ell({\rm\mathbf{X}},\mathbf{y})=-\sum_{i=1}^{B}\sum_{j=1}^{B}\sum_{k=1}^{K}\sum_{k^{\prime}=1}^{K}\mathbf{1}\{y_{j}=y_{i}\}\log\frac{\exp\psi(\mathbf{v}_{i}^{k},\mathbf{v}_{j}^{k^{\prime}})}{\sum_{j^{\dagger}=1}^{B}\sum_{k^{\dagger}=1}^{K}\mathbf{1}\{y_{j^{\dagger}}\neq y_{i}\}\exp\psi(\mathbf{v}_{i}^{k},\mathbf{v}_{j^{\dagger}}^{k^{\dagger}})}. (5)

𝐲={yi}iB\mathbf{y}=\{y_{i}\}_{i}^{B} are labels; 𝟏{}\mathbf{1}\{\cdot\} is indicator function. VPCL relieves multiple pretraining predictors required to adjust to different datasets. Moreover, VPCL exposes more feature embeddings to the supervision by partitioning hence providing more discriminative and generalizable representations.

Table 1: Statistics of the use clinical trial mortality prediction datasets. All are binary classification tasks. Positive ratio means the ratio of data points belong the positive class. NCTxxx are trial identifiers which can be linked to trials on ClinicalTrials.gov.
Name Datapoints Categorical Binary Numerical Positive ratio
NCT00041119 3871 5 8 2 0.07
NCT00174655 994 3 31 15 0.02
NCT00312208 1651 5 12 6 0.19
NCT00079274 2968 5 8 3 0.12
NCT00694382 1604 1 29 11 0.45

3 Experiments

In this section, we aim at answering the following questions by extensive experiments:

  • Q1. How does TransTab perform compared with baselines under the vanilla supervised setting?

  • Q2. How well does TransTab address incremental columns from a stream of data (S(2) in Fig. 1)?

  • Q3. How is the impact of TransTab learned from multiple tables (with different columns) drawn from the same domain on its predictive ability (S(1) in Fig. 1)?

  • Q4. Can TransTab be a zero-shot learner when pretrained on tables and infer on a new table (S(4) in Fig. 1)?

  • Q5. Is the proposed vertical partition CL better than vanilla supervised pretraining and self-supervised CL (S(3) in Fig. 1)?

Datasets. We introduce clinical trial mortality prediction datasets where each includes a distinct group of patients and columns 333https://data.projectdatasphere.org/projectdatasphere/html/access. The data statistics are in Table 1. Accurately predicting the patient mortality in clinical trials is crucial because it helps identify catastrophic treatment then save patients from harm and improve the clinical trial design. Considering they are from a similar domain, we can utilize them to test if TransTab can achieve transfer learning. Besides, we also include a set of public tabular datasets, the statistics are in Table 7.

Dataset pre-processing. For all baselines, we represent categorical features by ordinal encoding if they need to specify categorical features, otherwise one-hot encoding is used. Numerical features are scaled to [0,1][0,1] by min-max normalization. Exceptionally for TransTab, we map the categorical feature index to its original description, e.g., mapping class "1" under "gender" to "female".

Model and implementation protocols. Unless specified otherwise, we keep the settings fixed across all experiments. TransTab uses 2 layers of gated transformers where the embedding dimensions of numbers and tokens are 128, and the hidden dimension of intermediate dense layers is 256. The attention module has 8 heads. We choose ReLU activations and do not activate dropout. We train TransTab using Adam optimizer [27] with learning rate in {2e-5,5e-5,1e-4}\{2\text{e-5},5\text{e-5},1\text{e-4}\} and no weight decay; batch size is in {16,64,128}\{16,64,128\}. We set a maximum self-supervised pretraining epochs of 50 and supervised training epochs of 100. A patience of 10 is kept for supervised training for early stopping. Experiments were conducted with one RTX3070 GPU, i7-10700 CPU, and 16GB RAM.

Table 2: Test AUROC results on clinical trial mortality datasets the under supervised learning setting. All the remaining tables in this paper follow these setups to avoid clutter: the metric values are averaged over 10 random seeds; the Rank column reports the average rank across all datasets; Top results for each dataset are in bold.
Methods N00041119 N00174655 N00312208 N00079274 N00694382 Rank(Std)
LR 0.6364 0.8543 0.7382 0.7067 0.7360 5.40(1.14)
XGBoost 0.5937 0.5000 0.6911 0.6784 0.7440 9.60(3.71)
MLP 0.6340 0.6189 0.7427 0.6967 0.7063 8.00(2.83)
SNN 0.6335 0.9130 0.7469 0.6948 0.7246 5.80(2.39)
TabNet 0.5856 0.5401 0.6910 0.6031 0.7113 11.40(0.89)
DCN 0.6349 0.7577 0.7431 0.6952 0.7458 5.60(2.51)
AutoInt 0.6327 0.7502 0.7479 0.6958 0.7411 6.20(2.59)
TabTrans 0.6187 0.9035 0.7069 0.7178 0.7229 7.20(3.56)
FT-Trans 0.6372 0.9073 0.7586 0.7090 0.7231 4.20(2.28)
VIME 0.6397 0.8533 0.7227 0.6790 0.7232 7.00(3.08)
SCARF 0.6248 0.9310 0.7267 0.7176 0.7103 6.60(3.91)
TransTab 0.6408 0.9428 0.7770 0.7281 0.7648 1.00(0.00)
Table 3: Test AUROC results on clinical trial datasets under feature incremental learning.
Methods N00041119 N00174655 N00312208 N00079274 N00694382 Rank(Std)
LR 0.6213 0.8485 0.6801 0.6258 0.7236 4.6(3.21)
XGBoost 0.5735 0.7890 0.6760 0.6038 0.6463 8.8(2.59)
MLP 0.6371 0.7754 0.6871 0.6220 0.6851 6.2(2.95)
SNN 0.5765 0.7440 0.6854 0.6336 0.7035 6.4(2.30)
TabNet 0.5548 0.8419 0.5849 0.6052 0.6668 9.0(3.39)
DCN 0.5172 0.5846 0.6640 0.6535 0.6957 8.2(4.16)
AutoInt 0.5232 0.6075 0.7031 0.6394 0.6974 7.2(3.56)
TabTrans 0.5599 0.7652 0.6433 0.6365 0.6841 8.2(1.10)
FT-Trans 0.5552 0.8045 0.7148 0.6471 0.6815 5.8(3.11)
VIME 0.6101 0.8114 0.3705 0.6444 0.6436 7.4(4.22)
SCARF 0.5996 0.6261 0.7072 0.6535 0.6957 5.2(2.97)
TransTab 0.6797 0.8545 0.7617 0.6857 0.7795 1.0(0.00)

Baselines. We include the following baselines for comparison: Logistic regression (LR); XGBoost [28]; Multilayer perceptron (MLP); SeLU MLP (SNN) [29]; TabNet [30]; DCN [1]; AutoInt [31]; TabTransformer [5]; FT-Transformer [32]; VIME [2]; SCARF [11]. We provide the baseline architectures and implementations in Appendix B.

3.1 Q1. Supervised learning

Results of supervised learning on clinical trial mortality prediction datasets are summarized by Table 2. Note that all methods including ours do not perform pre-training. We see that our method outperforms baselines on all. From the view of method ranks, we surprisingly identify that LR wins over half of baseline methods. Except for TransTab, FT-Transformer is the only model that shows significant superiority over LR, which illustrates the potential of transformers for tabular modeling. Additional results on public datasets are available in Table 8 where we witness that our method is comparable to the state-of-the-art baseline tabular models. We also discover the baselines drawn from the CTR prediction literature (DCN and AutoInt) turn out the be competitive in tabular modeling.

3.2 Q2. Feature incremental learning

For previous tabular models, we should either drop new features or drop old data when confronting feature incremental learning. By contrast, TransTab is able to continually learn from new data with incremental features. We split the raw dataset into three subsets: set1, 2, and 3 which mimic the incremental feature scenario shown by (2) in Fig. 1. Baseline methods apply to two scenarios: (a) learning from all data that only have features of set1 and (b) learning from the data of set3 only. We report the best of the two. TransTab applies to learning from all three subsets. Table 3 shows the results where we find our method outperforms baselines by a great margin. It demonstrates that TransTab makes the best of incremental features to learn better. Similar observations appear in public datasets, shown by Table 9.

Table 4: Test AUROC results on clinical trial datasets under transfer learning across tables.
Methods N00041119 N00174655 N00312208 N00079274 N00694382 Rank(Std)
set1 set2 set1 set2 set1 set2 set1 set2 set1 set2
LR 0.625 0.647 0.789 0.819 0.701 0.735 0.635 0.685 0.675 0.763 5.33(1.73)
XGBoost 0.638 0.575 0.574 0.886 0.690 0.700 0.596 0.647 0.592 0.677 7.56(3.75)
MLP 0.639 0.621 0.314 0.857 0.683 0.744 0.620 0.675 0.648 0.765 6.56(3.32)
SNN 0.627 0.634 0.215 0.754 0.687 0.732 0.631 0.683 0.651 0.759 7.44(2.07)
TabNet 0.564 0.558 0.856 0.592 0.671 0.657 0.443 0.605 0.581 0.677 10.67(2.96)
DCN 0.636 0.625 0.767 0.790 0.711 0.698 0.682 0.664 0.658 0.737 6.33(2.45)
AutoInt 0.629 0.630 0.843 0.730 0.725 0.698 0.679 0.665 0.686 0.661 5.89(2.89)
TabTrans 0.616 0.647 0.866 0.822 0.675 0.677 0.618 0.702 0.652 0.718 6.22(3.38)
FT-Trans 0.627 0.641 0.836 0.858 0.720 0.741 0.692 0.692 0.652 0.740 4.22(2.28)
VIME 0.603 0.625 0.312 0.726 0.601 0.642 0.477 0.668 0.614 0.715 10.44(1.51)
SCARF 0.635 0.657 0.651 0.814 0.653 0.686 0.682 0.701 0.671 0.776 5.56(3.40)
TransTab 0.653 0.653 0.904 0.846 0.730 0.756 0.680 0.711 0.747 0.774 1.78(1.30)

3.3 Q3. Transfer learning

We further test if TransTab is able to transfer knowledge across tables. Results are in Table 4. We split each dataset into two subsets with 50% overlaps of their columns. Baselines are trained and tested on set1 (only label-supervision) or set2 separately. For our method we pretrain it on set1 then finetune it on set2 and report its performance on set2, and vice versa. We observe that TransTab can benefit from knowledge transfer across tables to reach superior performances. Similar observations are made on public datasets shown by Table 10.

Table 5: Test AUROC results on clinical trial datasets under zero-shot learning setting.
TransTab N00041119 N00174655 N00312208 N00079274 N00694382
Supervised 0.5854 0.6484 0.7536 0.7087 0.6479
Transfer 0.6130 0.6909 0.7658 0.7163 0.6752
Zero-shot 0.5990 0.6752 0.7576 0.7036 0.6740

3.4 Q4. Zero-shot learning

Although there are numerous papers on zero-shot learning (ZSL) in CV and NLP [33, 34, 35], we notice that ZSL was hardly mentioned in tabular domain. In this experiment, we refer to the ZSL scenario mentioned by S(4) of Fig. 1 where we split the raw table into three equal-size subsets. Three subsets have distinct columns. For the zero-shot setting, the model learns from set1+set2 and is tested on set3 without further training. In this scenario, the model needs to leverage the learned knowledge from set1 and set2 to support the inference on a new table set3. Besides, we design two baselines for comparison: supervised where the model learns from set3 and predicts on set3 and transfer where the model learns from set1+set2 and continues to be finetuned on set3. Results are in Table 5. We surprisingly find the ZSL model gets better performance than the supervised one on average. It boils down to that (1) ZSL TransTab succeeds to retain the learned knowledge from set1+set2 for predicting on a new table (set3) and (2) ZSL can benefit from more data (set1+set2) than the supervised (set3 only). Meanwhile, the transfer model takes the advantage of set1+set2 and is adapted for set3 by finetuning, hence reaches the best performance. Similarly, we witness that TransTab is able to make zero-shot predictions on public datasets as in Table 11.

Additional sensitivity check is provided by Fig. 6 where we vary the overlap ratio of two subsets from the same dataset. We witness that our model makes reasonable predictions even if the training set has no column overlap with the test set.

Table 6: Test AUROC on clinical trial datasets under the across-table pretraining plus finetuning setting. Supervised: baseline supervised model; Transfer: vanilla supervised transfer learning. Red shows the one worse than the Supervised baseline.
TransTab N00041119 N00174655 N00312208 N00079274 N00694382
Supervised 0.6313 0.8348 0.7444 0.6885 0.7293
Transfer 0.6424 0.8183 0.7458 0.6928 0.7239
Self-VPCL 0.6412 0.8577 0.7486 0.7069 0.7348
VPCL 0.6405 0.8583 0.7517 0.7063 0.7392

3.5 Q5. Supervised and self-supervised pretraining

We take experiments to compare the proposed VPCL with the vanilla transfer learning strategy, as in Table 6. We observe that the vanilla strategy harms the performance on two datasets while VPCL always brings positive effect for finetuning. Besides, we conduct experiments on varying the number of partitions and show the average AUROC on all five datasets, shown by Fig. 5. We specify that VPCL demonstrates an advantage over self-VPCL when we increase the partition numbers.

We also explore if pretraining works on public datasets. Results in Table 12 somewhat match our expectations that pretraining on unrelated tabular data usually yields few benefits for finetuning because these tables define totally different columns and targeted tasks. We also show the ablation on the number of partitions by Fig. 5 where VPCL consistently outperforms the Supervised baseline. Nevertheless, it is still worth investigating the table phenotypes to aggregate tables which are more likely to benefit from each other by transfer learning.

4 Related Works

Tabular Prediction. To enhance tabular predictions, numerous recent works try to design new algorithms [28, 36, 37, 30, 38, 32, 5, 10, 7, 39, 40, 41, 42, 43, 44, 45]. However, it was argued that boosting algorithms and MLPs are still the competitive choices for tabular data modeling, especially when the sample size is small [32, 46, 39, 47]. To alleviate label scarcity issue, SSL pretraining on unlabeled tabular data was introduced [2, 24, 10, 9, 11]. Nonetheless, none of them is transferable across tables then is able to extend the success of pretraining to the tabular domain. For practical tabular predictions, the common case is that we own a lot of labeled samples collected with diverse protocols hence heavy preprocessing is needed to align them by either dropping many samples or many features. By contrast, TransTab accepts variable-column tables and therefore can learn from different tables at scale and transfer to the target task. Also, it can support diverse tabular prediction tasks as depicted in Fig. 1, which cannot be done by off-the-shelf tabular methods.

Transfer learning. Transfer learning (TL) has long been a popular research field since the proposal of ImageNet [48], which gives rise to splendid works on utilizing supervised pretraining on a large general database and finetune on a small downstream task [49, 50, 51, 52, 53]. TL is also fast-growing in NLP beginning at BERT [20], which often leverages web-scale unlabeled texts for self-supervised pretraining and then applies to specific tasks [34, 54, 55, 56, 57]. However, few work was on TL in tabular predictions. As mentioned in §1, TransTab paves the way for effective tabular TL by establishing a feature processing protocol that applies for most table inputs, such that it shares knowledge across tables.

Self-supervised learning & contrastive learning. SSL uses unlabeled data with pretext tasks to learn useful representations and most of them are in CV and NLP [20, 17, 15, 16, 58, 23, 59, 60, 61, 62, 63]. Recent SSL tabular models can be classified into reconstruction and contrastive based methods: TabNet [30] and VIME [2] try to recover the corrupted inputs with auto-encoding loss; SCARF [11] takes a SimCLR-like [64] contrastive loss between the sample and its corrupted version; SubTab [9] takes a combination of both. Nevertheless, all fail to learn transferable models across tables such that cannot benefit from pretraining with scale. Contrastive learning can also be applied to supervised learning by leveraging class labels to build positive samples [26]. Our work extends it to to the tabular domain, which we prove works better than vanilla supervised pretraining. The vertical partition sampling also enjoys high query speed from large databases which are often column-oriented [25]. Another line of research takes table pretraining table semantic parsing [65, 66, 67, 68, 69] or table-to-text generation [70, 71]. But these methods either encode the whole table instead of each row or do not demonstrate to benefit tabular prediction yet.

5 Conclusion

This paper proposes TransTab that accepts variable-column inputs. By the proposed vertical partition contrastive learning, it can benefit from supervised pretraining from multiple tabular datasets with low memory cost. We envision it to be the basis of tabular foundation models and widely used to tabular-related applications in the future.

Acknowledgement

This work was supported by NSF award SCH-2205289, SCH-2014438, IIS-1838042, NIH award R01 1R01NS107291-01.

References

  • [1] Ruoxi Wang, Bin Fu, Gang Fu, and Mingliang Wang. Deep & cross network for ad click predictions. In Proceedings of the ADKDD’17, pages 1–7. 2017.
  • [2] Jinsung Yoon, Yao Zhang, James Jordon, and Mihaela van der Schaar. VIME: Extending the success of self-and semi-supervised learning to tabular domain. Advances in Neural Information Processing Systems, 33:11033–11043, 2020.
  • [3] Yixuan Zhang, Jialiang Tong, Ziyi Wang, and Fengqiang Gao. Customer transaction fraud detection using xgboost model. In International Conference on Computer Engineering and Application, pages 554–558. IEEE, 2020.
  • [4] Zifeng Wang and Suzhen Li. Data-driven risk assessment on urban pipeline network based on a cluster model. Reliability Engineering & System Safety, 196:106781, 2020.
  • [5] Xin Huang, Ashish Khetan, Milan Cvitkovic, and Zohar Karnin. TabTransformer: Tabular data modeling using contextual embeddings. arXiv preprint arXiv:2012.06678, 2020.
  • [6] Inkit Padhi, Yair Schiff, Igor Melnyk, Mattia Rigotti, Youssef Mroueh, Pierre Dognin, Jerret Ross, Ravi Nair, and Erik Altman. Tabular transformers for modeling multivariate time series. In IEEE International Conference on Acoustics, Speech and Signal Processing, pages 3565–3569. IEEE, 2021.
  • [7] Radostin Cholakov and Todor Kolev. The GatedTabTransformer. an enhanced deep learning architecture for tabular modeling. arXiv preprint arXiv:2201.00199, 2022.
  • [8] Zifeng Wang and Jimeng Sun. SurvTRACE: Transformers for survival analysis with competing events. arXiv preprint arXiv:2110.00855, 2021.
  • [9] Talip Ucar, Ehsan Hajiramezanali, and Lindsay Edwards. SubTab: Subsetting features of tabular data for self-supervised representation learning. Advances in Neural Information Processing Systems, 34, 2021.
  • [10] Gowthami Somepalli, Micah Goldblum, Avi Schwarzschild, C Bayan Bruss, and Tom Goldstein. SAINT: Improved neural networks for tabular data via row attention and contrastive pre-training. arXiv preprint arXiv:2106.01342, 2021.
  • [11] Dara Bahri, Heinrich Jiang, Yi Tay, and Donald Metzler. SCARF: Self-supervised contrastive learning using random feature corruption. In International Conference on Learning Representations, 2022.
  • [12] Fatemeh Nargesian, Erkang Zhu, Ken Q Pu, and Renée J Miller. Table union search on open data. Proceedings of the VLDB Endowment, 11(7):813–825, 2018.
  • [13] Erkang Zhu, Dong Deng, Fatemeh Nargesian, and Renée J Miller. Josie: Overlap set similarity search for finding joinable tables in data lakes. In International Conference on Management of Data, pages 847–864, 2019.
  • [14] Yuyang Dong, Kunihiro Takeoka, Chuan Xiao, and Masafumi Oyamada. Efficient joinable table discovery in data lakes: A high-dimensional similarity-based approach. In IEEE International Conference on Data Engineering, pages 456–467. IEEE, 2021.
  • [15] Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollár, and Ross Girshick. Masked autoencoders are scalable vision learners. arXiv preprint arXiv:2111.06377, 2021.
  • [16] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations, 2020.
  • [17] Hangbo Bao, Li Dong, and Furu Wei. BEiT: Bert pre-training of image transformers. arXiv preprint arXiv:2106.08254, 2021.
  • [18] Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781, 2013.
  • [19] Mike Schuster and Kaisuke Nakajima. Japanese and korean voice search. In IEEE international conference on acoustics, speech and signal processing (ICASSP), pages 5149–5152. IEEE, 2012.
  • [20] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  • [21] Bill Yuchen Lin, Seyeon Lee, Rahul Khanna, and Xiang Ren. Birds have four legs?! NumerSense: Probing numerical commonsense knowledge of pre-trained language models. In Proceedings of the EMNLP, pages 6862–6868, 2020.
  • [22] Jimmy Lei Ba, Jamie Ryan Kiros, and Geoffrey E Hinton. Layer normalization. arXiv preprint arXiv:1607.06450, 2016.
  • [23] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in Neural Information Processing Systems, 30, 2017.
  • [24] Sajad Darabi, Shayan Fazeli, Ali Pazoki, Sriram Sankararaman, and Majid Sarrafzadeh. Contrastive mixup: Self-and semi-supervised learning for tabular domain. arXiv preprint arXiv:2108.12296, 2021.
  • [25] Mike Stonebraker, Daniel J Abadi, Adam Batkin, Xuedong Chen, Mitch Cherniack, Miguel Ferreira, Edmond Lau, Amerson Lin, Sam Madden, Elizabeth O’Neil, et al. C-store: a column-oriented dbms. In Making Databases Work: the Pragmatic Wisdom of Michael Stonebraker, pages 491–518. 2018.
  • [26] Prannay Khosla, Piotr Teterwak, Chen Wang, Aaron Sarna, Yonglong Tian, Phillip Isola, Aaron Maschinot, Ce Liu, and Dilip Krishnan. Supervised contrastive learning. Advances in Neural Information Processing Systems, 33:18661–18673, 2020.
  • [27] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  • [28] Tianqi Chen and Carlos Guestrin. Xgboost: A scalable tree boosting system. In ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 785–794, 2016.
  • [29] Günter Klambauer, Thomas Unterthiner, Andreas Mayr, and Sepp Hochreiter. Self-normalizing neural networks. Advances in Neural Information Processing Systems, 30, 2017.
  • [30] Sercan O Arık and Tomas Pfister. Tabnet: Attentive interpretable tabular learning. In AAAI, volume 35, pages 6679–6687, 2021.
  • [31] Weiping Song, Chence Shi, Zhiping Xiao, Zhijian Duan, Yewen Xu, Ming Zhang, and Jian Tang. AutoInt: Automatic feature interaction learning via self-attentive neural networks. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, pages 1161–1170, 2019.
  • [32] Yury Gorishniy, Ivan Rubachev, Valentin Khrulkov, and Artem Babenko. Revisiting deep learning models for tabular data. Advances in Neural Information Processing Systems, 34, 2021.
  • [33] Bernardino Romera-Paredes and Philip Torr. An embarrassingly simple approach to zero-shot learning. In International conference on machine learning, pages 2152–2161. PMLR, 2015.
  • [34] Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are few-shot learners. Advances in Neural Information Processing Systems, 33:1877–1901, 2020.
  • [35] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International Conference on Machine Learning, pages 8748–8763. PMLR, 2021.
  • [36] Guolin Ke, Qi Meng, Thomas Finley, Taifeng Wang, Wei Chen, Weidong Ma, Qiwei Ye, and Tie-Yan Liu. LightGBM: A highly efficient gradient boosting decision tree. Advances in Neural Information Processing Systems, 30, 2017.
  • [37] Anna Veronika Dorogush, Vasily Ershov, and Andrey Gulin. CatBoost: gradient boosting with categorical features support. arXiv preprint arXiv:1810.11363, 2018.
  • [38] Jintai Chen, Kuanlun Liao, Yao Wan, Danny Z Chen, and Jian Wu. Danets: Deep abstract networks for tabular data classification and regression. arXiv preprint arXiv:2112.02962, 2021.
  • [39] Vadim Borisov, Tobias Leemann, Kathrin Seßler, Johannes Haug, Martin Pawelczyk, and Gjergji Kasneci. Deep neural networks and tabular data: A survey. arXiv preprint arXiv:2110.01889, 2021.
  • [40] Ami Abutbul, Gal Elidan, Liran Katzir, and Ran El-Yaniv. Dnf-net: A neural architecture for tabular data. arXiv preprint arXiv:2006.06465, 2020.
  • [41] Liran Katzir, Gal Elidan, and Ran El-Yaniv. Net-dnf: Effective deep modeling of tabular data. In International Conference on Learning Representations, 2020.
  • [42] Yuanfei Luo, Hao Zhou, Wei-Wei Tu, Yuqiang Chen, Wenyuan Dai, and Qiang Yang. Network on network for tabular data classification in real-world applications. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 2317–2326, 2020.
  • [43] Xiawei Guo, Yuhan Quan, Huan Zhao, Quanming Yao, Yong Li, and Weiwei Tu. TabGNN: Multiplex graph neural network for tabular data prediction. arXiv preprint arXiv:2108.09127, 2021.
  • [44] Yuanfei Luo, Mengshuo Wang, Hao Zhou, Quanming Yao, Wei-Wei Tu, Yuqiang Chen, Wenyuan Dai, and Qiang Yang. Autocross: Automatic feature crossing for tabular data in real-world applications. In ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 1936–1945, 2019.
  • [45] Jiarui Qin, Weinan Zhang, Rong Su, Zhirong Liu, Weiwen Liu, Ruiming Tang, Xiuqiang He, and Yong Yu. Retrieval & interaction machine for tabular data prediction. In ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pages 1379–1389, 2021.
  • [46] Ravid Shwartz-Ziv and Amitai Armon. Tabular data: Deep learning is not all you need. Information Fusion, 81:84–90, 2022.
  • [47] Arlind Kadra, Marius Lindauer, Frank Hutter, and Josif Grabocka. Well-tuned simple nets excel on tabular datasets. Advances in Neural Information Processing Systems, 34, 2021.
  • [48] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In IEEE Conference on Computer Vision and Pattern Recognition, pages 248–255. IEEE, 2009.
  • [49] Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556, 2014.
  • [50] Jason Yosinski, Jeff Clune, Yoshua Bengio, and Hod Lipson. How transferable are features in deep neural networks? Advances in Neural Information Processing Systems, 27, 2014.
  • [51] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In IEEE Conference on Computer Vision and Pattern Recognition, pages 770–778, 2016.
  • [52] Gao Huang, Zhuang Liu, Laurens Van Der Maaten, and Kilian Q Weinberger. Densely connected convolutional networks. In IEEE Conference on Computer Vision and Pattern Recognition, pages 4700–4708, 2017.
  • [53] Barret Zoph, Vijay Vasudevan, Jonathon Shlens, and Quoc V Le. Learning transferable architectures for scalable image recognition. In IEEE Conference on Computer Vision and Pattern Recognition, pages 8697–8710, 2018.
  • [54] Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692, 2019.
  • [55] Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J Liu. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research, 21:1–67, 2020.
  • [56] Zhilin Yang, Zihang Dai, Yiming Yang, Jaime Carbonell, Russ R Salakhutdinov, and Quoc V Le. XLNet: Generalized autoregressive pretraining for language understanding. Advances in Neural Information Processing Systems, 32, 2019.
  • [57] Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Veselin Stoyanov, and Luke Zettlemoyer. BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Annual Meeting of the Association for Computational Linguistics, pages 7871–7880, 2020.
  • [58] Tianyu Gao, Xingcheng Yao, and Danqi Chen. Simcse: Simple contrastive learning of sentence embeddings. In Conference on Empirical Methods in Natural Language Processing, pages 6894–6910, 2021.
  • [59] Jacob Devlin Ming-Wei Chang Kenton and Lee Kristina Toutanova. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of NAACL-HLT, pages 4171–4186, 2019.
  • [60] Yao-Hung Hubert Tsai, Shaojie Bai, Paul Pu Liang, J Zico Kolter, Louis-Philippe Morency, and Ruslan Salakhutdinov. Multimodal transformer for unaligned multimodal language sequences. In Proceedings of the Annual Meeting of the Association for Computational Linguistics, 2019.
  • [61] Hassan Akbari, Liangzhe Yuan, Rui Qian, Wei-Hong Chuang, Shih-Fu Chang, Yin Cui, and Boqing Gong. VATT: Transformers for multimodal self-supervised learning from raw video, audio and text. Advances in Neural Information Processing Systems, 34, 2021.
  • [62] Dmitry Lepikhin, HyoukJoong Lee, Yuanzhong Xu, Dehao Chen, Orhan Firat, Yanping Huang, Maxim Krikun, Noam Shazeer, and Zhifeng Chen. GShard: Scaling giant models with conditional computation and automatic sharding. In International Conference on Learning Representations, 2020.
  • [63] William Fedus, Barret Zoph, and Noam Shazeer. Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity. arXiv preprint arXiv:2101.03961, 2021.
  • [64] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pages 1597–1607. PMLR, 2020.
  • [65] Pengcheng Yin, Graham Neubig, Wen-tau Yih, and Sebastian Riedel. TaBERT: Pretraining for joint understanding of textual and tabular data. In Annual Meeting of the Association for Computational Linguistics, pages 8413–8426, 2020.
  • [66] Hiroshi Iida, Dung Thai, Varun Manjunatha, and Mohit Iyyer. TABBIE: Pretrained representations of tabular data. In Conference of the North American Chapter of the Association for Computational Linguistics, 2021.
  • [67] Xi Victoria Lin, Richard Socher, and Caiming Xiong. Bridging textual and tabular data for cross-domain text-to-sql semantic parsing. arXiv preprint arXiv:2012.12627, 2020.
  • [68] Xiang Deng, Huan Sun, Alyssa Lees, You Wu, and Cong Yu. TURL: Table understanding through representation learning. Proceedings of the VLDB Endowment, 2021.
  • [69] Jingfeng Yang, Aditya Gupta, Shyam Upadhyay, Luheng He, Rahul Goel, and Shachi Paul. TableFormer: Robust transformer modeling for table-text encoding. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, pages 528–537, 2022.
  • [70] Fei Wang, Kexuan Sun, Muhao Chen, Jay Pujara, and Pedro A Szekely. Retrieving complex tables with multi-granular graph representation learning. In SIGIR, 2021.
  • [71] Fei Wang, Zhewei Xu, Pedro Szekely, and Muhao Chen. Robust (controlled) table-to-text generation with structure-aware equivariance learning. arXiv preprint arXiv:2205.03972, 2022.
  • [72] Vincenzo Cutrona. Semantic enrichment for large-scale data analytics. 2019.
  • [73] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. Advances in Neural Information Processing Systems, 32, 2019.
  • [74] Aaron Van den Oord, Yazhe Li, Oriol Vinyals, et al. Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748, 2(3):4, 2018.

Checklist

  1. 1.

    For all authors…

    1. (a)

      Do the main claims made in the abstract and introduction accurately reflect the paper’s contributions and scope? [Yes]

    2. (b)

      Did you describe the limitations of your work? [Yes] See § 3.5 and Appendix §A.

    3. (c)

      Did you discuss any potential negative societal impacts of your work? [Yes] See Appendix §A.

    4. (d)

      Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes]

  2. 2.

    If you are including theoretical results…

    1. (a)

      Did you state the full set of assumptions of all theoretical results? [N/A] This paper does not include theoretical results.

    2. (b)

      Did you include complete proofs of all theoretical results? [N/A] This paper does not include theoretical results.

  3. 3.

    If you ran experiments…

    1. (a)

      Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] See supplementary materials.

    2. (b)

      Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] For our methods please see Model and Implementation Protocols of §3; for the compared baselines please see Appendix §B;

    3. (c)

      Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes]

    4. (d)

      Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] Please see Model and Implementation Protocols of §3.

  4. 4.

    If you are using existing assets (e.g., code, data, models) or curating/releasing new assets…

    1. (a)

      If your work uses existing assets, did you cite the creators? [Yes] See Table 13.

    2. (b)

      Did you mention the license of the assets? [Yes] Licenses are available referring to the provided links.

    3. (c)

      Did you include any new assets either in the supplemental material or as a URL? [Yes] See Appendix §C.

    4. (d)

      Did you discuss whether and how consent was obtained from people whose data you’re using/curating? [Yes] See Appendix §C.

    5. (e)

      Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [Yes] See Appendix §C.

  5. 5.

    If you used crowdsourcing or conducted research with human subjects…

    1. (a)

      Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] This paper does not use crowdsourcing.

    2. (b)

      Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] This paper does not use crowdsourcing.

    3. (c)

      Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A] This paper does not use crowdsourcing.

Appendix A Broader impact of this work

Tabular data is the most common data format used in practical data analysis including healthcare, finance, business, advertising, manufacturing, etc. Regardless of its importance, much more effort was made to play deep learning with other domains like vision, language, and audio. As a result, non-deep algorithms especially tree-based still dominate tabular data analysis. This paper opens the door to leverage the power of transfer learning in the tabular domain because TransTab is capable of dealing with variable-column input tables. This property also alleviates the workload required for data preprocessing because TransTab permits missing features and make good predictions based on the remaining features. This advance is expected to bring tremendous savings of time and money from the data engineering which often takes up 80% of efforts for data science projects [72]. It also sheds light on developing foundation models for the tabular domain because of the success of pretraining on scale.

The potential negative effect would be that it requires more effort on solving data privacy issues because TransTab works better when features are represented by descriptive texts compared with discretized indices. However, this can be alleviated by mapping features to machine-readable tokens in advance by a private codebook. Besides, TransTab needs more resources than simple models like MLP and Trees. The main cause is the use of full attention in multihead attention modules and the tokenization in the featurizing process. The former problem can be alleviated by replacing transformers with MLP-based blocks like gated attention units. The latter can be circumvented by pre-tokenization before the model training.

Appendix B Baseline architecture and implementation

In this section, we introduce the implementation details of baselines in experiments.

  • LR: Use the default setting of the package scikit-learn444sklearn.linear_model.LogisticRegression except the max iterations is set 1000.

  • XGBoost: We set the maximum number of estimators in {50,100,500}\{50,100,500\} and the max depth in {4,6,8}\{4,6,8\}. Implemented based on the XGBoost package555xgboost.sklearn.

  • MLP & SNN: We keep the same architecture for both except for their activations: three dense layers with hidden dimensions 256, 256, 1; dropout with rate of 0.1 is used. They are trained with batch size {16,32,64,128}\in\{16,32,64,128\}, learning rate \in {5e-5, 1e-4, 5e-4, 1e-3}, and early stopping patience of 5 with 100 maximum epochs.

  • TabNet: Use the official implementation with the default recommended parameters666https://github.com/dreamquark-ai/tabnet. Trained with batch size {16,32,64,128}\in\{16,32,64,128\}, learning rate \in {1e-4, 1e-3, 2e-2}, na,nd{8,16,64,128}n_{a},n_{d}\in\{8,16,64,128\}, γ{1.3,1.5,1.8}\gamma\in\{1.3,1.5,1.8\}, categorical embedding dimension {1,8,16}\in\{1,8,16\} and early stopping patience of 5 with 100 maximum epochs.

  • DCN: Use the implementation by Deep-CTR777https://github.com/shenweichen/DeepCTR-Torch. The number of cross is 2; dropout rate for feed-forward component is 0.1; MLP part has two dense layers of dimension 256, 128; Trained with batch size {16,32,64,128}\in\{16,32,64,128\}, learning rate \in {5e-5, 1e-4, 1e-3}, and early stopping patience of 10 in 100 maximum epochs.

  • AutoInt: Use the implementation by Deep-CTR. Attention layer number is set 2; Attention head number is set 2; MLP part has two dense layers of dimension 256, 128; dropout deactivated; Trained with batch size {16,32,64,128}\in\{16,32,64,128\}, learning rate \in {5e-5, 1e-4, 1e-3}, and early stopping patience of 10 in 100 maximum epochs.

  • TabTransformer: Use the official implementation888https://github.com/lucidrains/tab-transformer-pytorch. Feed-forward component has 128 dimension; 2 transformer layers are used; The number of heads of attention is {2,4,8}\in\{2,4,8\}; Dropout rate is 0.1; ReLU activation is used; Trained with batch size {16,32,64,128}\in\{16,32,64,128\}, learning rate \in {5e-5,1e-4,1e-3}, and early stopping patience of 10 in 100 maximum epochs.

  • FT-Transformer: Use the official implementation999https://github.com/Yura52/rtdl. Feed-forward component has 128 dimension; 2 transformer layers are used; The number of heads of attention is {2,4,8}\in\{2,4,8\}; Dropout rate is 0.1; ReLU activation is used; Trained with batch size {16,32,64}\in\{16,32,64\}, learning rate \in {5e-5,1e-4,1e-3}, and early stopping patience of 10 in 100 maximum epochs.

  • VIME: We reproduce it by PyTorch [73] based on the original official implementation101010https://github.com/jsyoon0823/VIME where its encoders, mask estimators, decoders are all one dense layer with the hidden dimension same as the input features. During the pretraining phase, we train the model on all training data taking mask rate 0.3, batch size 128, learning rate 1e-4, and 10 epochs; during the fine-tuning phase, we add a classifier after the encoder with three dense layers of 100 dimension and ReLU activations. Trained with batch size {16,32,64,128}\in\{16,32,64,128\}, learning rate {5e-5,1e-4,1e-3}, and early stopping patience of 10 in 100 maximum epochs.

  • SCARF: Since no official code is found, we reproduce it by PyTorch based on the descriptions in the original paper. A 4-layer encoder and a 2-layer decoder with ReLU activations are used. Hidden dimensions of their intermediate layers are all 256. During the pretraining phase, we train the model on all trainign data taking mask rate 0.5, InfoNCE loss [74] with learning rate 1e-3, batch size 128, and 10 epochs; during the fine-tuning phase, we add a classifier after the encoder with two dense layers with 256 hidden dimensions. Trained with batch size {16,32,64,128}\in\{16,32,64,128\}, learning rate \in {5e-5,1e-4,1e-3}, and early stopping patience of 10 in 100 maximum epochs.

Table 7: Statistics of the used public datasets. All are binary classification tasks. Positive ratio means the ratio of data points belong the positive class. Source links are available at Table 13.
Name N Datapoints Categorical Binary Numerical Positive ratio
credit-g (CG) 1000 11 2 7 0.7
credit-approval (CA) 690 6 3 6 0.56
dresses-sales (DS) 500 11 0 1 0.42
adult (AD) 48842 12 0 2 0.24
cylinder-bands (CB) 540 13 4 18 0.58
blastchar (BL) 7043 11 5 3 0.27
insurance-co (IO) 5822 2 0 83 0.06
1995-income (IC) 32561 8 0 6 0.24
Table 8: Test AUROC results on public datasets the under supervised learning setting. The dataset name is abbreviated referring to Table 7.
Methods CG CA DS AD CB BL IO IC Rank(Std)
LR 0.720 0.836 0.557 0.851 0.748 0.801 0.769 0.860 9.88(1.90)
XGBoost 0.726 0.895 0.587 0.912 0.892 0.821 0.758 0.925 5.12(3.86)
MLP 0.643 0.832 0.568 0.904 0.613 0.832 0.779 0.893 9.25(2.07)
SNN 0.641 0.880 0.540 0.902 0.621 0.834 0.794 0.892 8.00(3.32)
TabNet 0.585 0.800 0.478 0.904 0.680 0.819 0.742 0.896 10.75(1.49)
DCN 0.739 0.870 0.674 0.913 0.848 0.840 0.768 0.915 4.12(1.69)
AutoInt 0.744 0.866 0.672 0.913 0.808 0.844 0.762 0.916 4.62(2.52)
TabTrans 0.718 0.860 0.648 0.914 0.855 0.820 0.794 0.882 6.50(3.12)
FT-Trans 0.739 0.859 0.657 0.913 0.862 0.841 0.793 0.915 3.94(1.35)
VIME 0.735 0.852 0.485 0.912 0.769 0.837 0.786 0.908 6.06(2.83)
SCARF 0.733 0.861 0.663 0.911 0.719 0.833 0.758 0.905 6.75(2.12)
TransTab 0.768 0.881 0.643 0.907 0.851 0.845 0.822 0.919 3.00(1.93)
Table 9: Test AUROC results on public datasets the under feature incremental learning setting. The dataset name is abbreviated referring to Table 7.
Methods CG CA DS AD CB BL IO IC Rank(Std)
LR 0.670 0.773 0.475 0.832 0.727 0.806 0.655 0.825 8.88(2.80)
XGBoost 0.608 0.817 0.527 0.891 0.778 0.816 0.692 0.898 5.50(3.13)
MLP 0.586 0.676 0.516 0.890 0.631 0.825 0.626 0.885 9.50(2.07)
SNN 0.583 0.738 0.442 0.888 0.644 0.818 0.643 0.881 9.69(1.28)
TabNet 0.573 0.689 0.419 0.886 0.571 0.837 0.680 0.882 9.19(3.34)
DCN 0.674 0.835 0.578 0.893 0.778 0.840 0.660 0.891 3.38(2.01)
AutoInt 0.671 0.825 0.563 0.893 0.769 0.836 0.676 0.887 4.75(1.25)
TabTrans 0.653 0.732 0.584 0.856 0.784 0.792 0.674 0.828 7.62(3.78)
FT-Trans 0.662 0.824 0.626 0.892 0.768 0.840 0.645 0.889 4.81(2.59)
VIME 0.621 0.697 0.571 0.892 0.769 0.803 0.683 0.881 7.00(3.01)
SCARF 0.651 0.753 0.556 0.891 0.703 0.829 0.680 0.887 6.56(1.32)
TransTab 0.741 0.879 0.665 0.894 0.791 0.841 0.739 0.897 1.12(0.35)
Table 10: Test AUROC results on public datasets under transfer learning across tables.
Methods CG CA DS AD CB BL IO IC Rank(Std)
set1 set2 set1 set2 set1 set2 set1 set2 set1 set2 set1 set2 set1 set2 set1 set2
LR 0.69 0.69 0.81 0.82 0.47 0.56 0.81 0.81 0.68 0.78 0.77 0.82 0.71 0.81 0.81 0.84 8.57(2.83)
XGB 0.72 0.71 0.85 0.87 0.46 0.63 0.88 0.89 0.80 0.81 0.76 0.82 0.65 0.74 0.92 0.91 5.57(3.62)
MLP 0.67 0.70 0.82 0.86 0.53 0.67 0.89 0.90 0.73 0.82 0.79 0.83 0.70 0.78 0.90 0.90 5.00(2.86)
SNN 0.66 0.63 0.85 0.83 0.54 0.42 0.87 0.88 0.57 0.54 0.77 0.82 0.69 0.78 0.87 0.88 8.67(2.92)
TabNet 0.60 0.47 0.66 0.68 0.54 0.53 0.87 0.88 0.58 0.62 0.75 0.83 0.62 0.71 0.88 0.89 9.87(2.92)
DCN 0.69 0.70 0.83 0.85 0.51 0.58 0.88 0.74 0.79 0.78 0.79 0.76 0.70 0.71 0.91 0.90 6.67(2.94)
AutoInt 0.70 0.70 0.82 0.86 0.49 0.55 0.88 0.74 0.77 0.79 0.79 0.76 0.71 0.72 0.91 0.90 6.03(3.23)
TabTrans 0.72 0.72 0.84 0.86 0.54 0.57 0.88 0.90 0.73 0.79 0.78 0.81 0.67 0.71 0.88 0.88 6.03(2.93)
FT-Trans 0.72 0.71 0.83 0.85 0.53 0.64 0.89 0.90 0.76 0.79 0.78 0.84 0.68 0.78 0.91 0.91 4.97(1.95)
VIME 0.59 0.70 0.79 0.76 0.45 0.53 0.88 0.90 0.65 0.81 0.58 0.83 0.67 0.70 0.90 0.90 8.83(3.24)
SCARF 0.69 0.72 0.82 0.85 0.55 0.64 0.88 0.89 0.77 0.73 0.78 0.83 0.71 0.75 0.90 0.89 5.47(2.42)
TransTab 0.74 0.76 0.87 0.89 0.55 0.66 0.88 0.90 0.80 0.80 0.79 0.84 0.73 0.82 0.91 0.91 2.33(2.10)
Table 11: Test AUROC results on public datasets under zero-shot learning setting.
TransTab CG CA DS AD CB BL IO IC
Supervised 0.581 0.635 0.571 0.898 0.733 0.822 0.702 0.875
Transfer 0.719 0.758 0.561 0.900 0.854 0.831 0.761 0.880
Zero-shot 0.685 0.721 0.538 0.892 0.710 0.804 0.742 0.874
Table 12: Test AUROC on public datasets under the across-table pretraining plus finetuning setting. Supervised: baseline supervised model; Transfer: vanilla supervised transfer learning. Red shows the one worse than the baseline Supervised.
TransTab CG CA DS AD CB BL IO IC
Supervised 0.763 0.858 0.630 0.907 0.841 0.844 0.821 0.919
Transfer 0.786 0.861 0.653 0.907 0.819 0.843 0.813 0.918
Self-VPCL 0.777 0.837 0.626 0.907 0.819 0.843 0.823 0.919
VPCL 0.776 0.858 0.637 0.907 0.862 0.844 0.819 0.919
Refer to caption
Figure 4: Analysis of the number of partitions for VPCL and self-VPCL on the clinical trial datasets.
Refer to caption
Figure 5: Analysis of the number of partitions for VPCL and self-VPCL on the public datasets
Table 13: Benchmark dataset links.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 6: Evaluate how the overlap ratio of two tables’ columns influences zero-shot learning (ZSL) performance of TransTab, on public data CG, CA, CB, and DS. xx-axis: the ratio of test table columns exist in the training table (0: no test table column appears in training table; 1: all test table columns in training table); yy-axis: the test AUC when making ZSL.

Appendix C Preprocessing of clinical trial datasets

The raw real-world de-identified patient-level clinical records are obtained from Project Data Sphere. All clinical trial data used in this paper are available under registration for the data platform. For each trial, we manage to extract patient’s baseline information as the features, including demographic information, medical history, medication history, lab test, vital signs, adverse events, etc. We draw the target labels from the survival analysis section where censoring is considered as "alive" and other events are tagged "mortality" so as to transform the datasets into binary prediction tasks.

Appendix D Establishment of subsets

For experiments in §3.2, §3.3, §3.4, we create subsets randomly with a fixed seed, respectively. That is, subsets vary across these experiments.

  • Feature incremental learning. The columns are splitted into three distinct parts v1,v2,v3{v_{1},v_{2},v_{3}}. Set1 contains v1v_{1}, set2 contains v1,v2v_{1},v_{2}, and set3 has v1,v2,v3v_{1},v_{2},v_{3}. Three sets have the equal number of samples.

  • Transfer learning. The columns are splitted into two parts v1,v2v_{1},v_{2} where v1v_{1} and v2v_{2} have 50% of elements overlapped. Two sets have the equal number of samples.

  • Zeroshot learning. The columns are splitted into three distinct parts v1,v2,v3{v_{1},v_{2},v_{3}}. Set1 contains v1v_{1}, set2 contains v2v_{2}, set3 contains v3v_{3}. Three sets have the equal number of samples.