Enhanced Exploration in Neural Feature Selection for Deep Click-Through Rate Prediction Models via Ensemble of Gating Layers
Abstract
Feature selection has been an essential step in developing industry-scale deep Click-Through Rate (CTR) prediction systems. The goal of neural feature selection (NFS) is to choose a relatively small subset of features with the best explanatory power as a means to remove redundant features and reduce computational cost. Inspired by gradient-based neural architecture search (NAS) and network pruning methods, people have tackled the NFS problem with Gating approach that inserts a set of differentiable binary gates to drop less informative features. The binary gates are optimized along with the network parameters in an efficient end-to-end manner. In this paper, we analyze the gradient-based solution from an exploration-exploitation perspective and use empirical results to show that Gating approach might suffer from insufficient exploration. To improve the exploration capacity of gradient-based solution, we propose a simple but effective ensemble learning approach, named Ensemble Gating. We choose two public datasets, namely Avazu and Criteo, to evaluate this approach. Our experiments show that, without adding any computational overhead or introducing any hyper-parameter (except the size of the ensemble), our method is able to consistently improve Gating approach and find a better subset of features on the two datasets with three different underlying deep CTR prediction models.
Introduction
Deep Click-Through Rate (CTR) prediction systems have shown promising performance in many industry-scale CTR prediction tasks, but this usually comes with high computational cost and high memory usage. To make deep learning models more memory efficient, people have proposed many approaches to reduce the size of the networks, such as network pruning (Xia, Zigeng, and Sanguthevar 2019; Gao et al. 2020), training with resource constraint (Srinivas, Subramanya, and Venkatesh Babu 2017), and adjusting feature embedding sizes (Joglekar et al. 2019; Liu et al. 2021). In this work, we focus on improving memory efficiency by performing neural feature selection that chooses a relatively small subset of features with good explanatory power.
We consider end-to-end feature selection approaches, which interpret the feature selection problem as a differentiable neural architecture search (NAS) or network pruning problem, and thus enable people to use many well-studied methods from the two relevant domains. In those approaches, to remove less predictive features, a set of “binary” gates with trainable weights are added to the network to indicate if certain features should remain or drop (Fig. 1). Depending on how discrete relaxation is achieved, there can be different implementations of the binary gates (Sheth and Fusi 2020; Yamada et al. 2020). Similar to differentiable NAS (Han, Ligeng, and Song 2019) and network pruning (Xia, Zigeng, and Sanguthevar 2019; Gao et al. 2020), differentiable neural feature selection methods realize their functionality by iteratively performing feature pruning (updating binary gates) and network fine-tuning (updating network parameters) over a pretrained over-parameterized network. In this paper, we refer to the methods that use such a gating formula as Gating method and the gate output as gating decision.
Since the feature selection problem is essentially a search problem with the search space comprised of all possible gating decisions, to find the most informative subset of features, the gates must fully explore the search space by taking various gating decisions and comparing the outcomes. However, in the Gating approach, the gating exploration is mainly driven by the gradient signal, which doesn’t always provide a strong enough exploration incentive, especially when the binary gates are jointly optimized with the model parameters. Here, we present a special experiment to illustrate how insufficient exploration can result in undesirable outcomes. In this experiment, we use a step function and straight through estimator (Bengio, Léonard, and Courville 2013; Hubara et al. 2016) to implement the binary gates. Details of the implementations and the searching/training process can be found in later sections. But different from normal settings, we deliberately reduce the learning rate of binary gates by a factor of ten. The goal of this setting change is to exaggerate the negative effects of insufficient exploration. Specifically, suppose the gradient signal suggests the gate should be changed from “on” to “off” and suppose there is no other exploration incentive except the gradient, since the learning rate of binary gates is abnormally small, it takes more update steps for the gate to change its state. However, before the gate is turned off, the model parameters have already been fine-tuned according to current gating decision. In this case, changing gate states will instead increase the training loss. We call this phenomenon gating overfitting. The experiment result verifies our hypothesis (Fig. 2). Note that gating overfitting exists even when all the hyper-parameters are properly set.


To enhance gating exploration and prevent gating overfitting, we propose a scalable and effective ensemble learning method, named Ensemble Gating, which creates multiple groups of binary gates and randomly selects one group to perform parameters update at each step. We show that, without adding any computational overhead or introducing any hyper-parameter (except the size of the ensemble), Ensemble Gating can provide meaningful uncertainty-driven exploration. We conduct extensive experiments on two public datasets, namely Avazu and Criteo. The results demonstrate that Ensemble Gating consistently finds better subsets of features on both datasets for three different deep CTR prediction models. Further studies show that Ensemble Gating converges quickly within a few epochs, and the uncertainty-driven exploration is more effective than exploring by injecting random noise into the learning system.
Related Works
The idea of using differentiable binary gates to automatically select important features was discussed in (Sheth and Fusi 2020; Yamada et al. 2020). It has also been widely applied in industry and has been studied in many relevant research directions like gradient-based neural architecture search (Han, Ligeng, and Song 2019; Liu, Simonyan, and Yang 2018) and network pruning (Han et al. 2015; Srinivas, Subramanya, and Venkatesh Babu 2017; Ye et al. 2020), which use learnable binary gates to determine whether to prune certain weight or neuron. Gating approaches in neural feature selection can be viewed as a reduced version of related network pruning approaches in the sense that binary gates are only inserted between certain layers in neural feature selection.
Due to the discrete nature of feature selection/network pruning problems (there are only two possible actions, namely ”prune” or ”keep”), there is no direct way to optimize the binary gates along with the deep neural network in a fully end-to-end manner. Thus, substantial research efforts have been focusing on finding better approximation that relaxes the discrete search space to a continuous space by using some differentiable re-parameterization tricks. Works like (Xia, Zigeng, and Sanguthevar 2019) and (Gao et al. 2020) use straight through estimator (Bengio, Léonard, and Courville 2013; Hubara et al. 2016) to enable gradients to be propagated. Some other works also use Gumbel-Softmax (Jang, Gu, and Poole 2016; Maddison, Mnih, and Teh 2016) to do the discrete relaxation (Xie et al. 2018; Liu, Simonyan, and Yang 2018).
The problem of neural feature selection (NFS) is also related to the neural input search (NIS) problem (Joglekar et al. 2019; Ginart et al. 2019). But the latter aims to reduce memory usage by assigning varying embedding sizes to different features rather than dropping an entire feature. Many works employ a similar differentiable gating formula to tackle the NIS problem (Zhao et al. 2021; Cheng, Shen, and Huang 2020; Liu et al. 2021), but in their frameworks, each binary gate is used to control the dimension of each embedding vector. Note that NIS and NFS are not interchangeable. Firstly, NIS can not be applied to models that require feature embeddings of the same size, such as DeepFM (Guo et al. 2017). Secondly, it can be more effective to perform feature selection before applying NIS or other network pruning techniques, especially for industry-scale CTR prediction models.
Feature selection for CTR prediction systems has been explored previously (Ronen et al. 2013; Koenigstein and Paquet 2013). In this work, we focus on improving the gradient-based feature methods, which are shown to be more computationally efficient with large-scale deep learning models (Sheth and Fusi 2020). Comprehensive surveys on feature selection are provided by Parmezan et al. and Chandrashekar and Sahin.
Preliminaries
Figure 1 depicts a typical deep learning CTR prediction model. We assume the model input consists of categorical feature fields. Each raw categorical feature is initially represented by a sparse one-hot vector. Then the feature embedding layer transforms each sparsely encoded feature into a -dimensional embedding vector as follows:
(1) |
where denotes the embedding matrix, and denotes the size of sparse encoding. Note that in this paper, we assume a global embedding size (default value: 8) is used for all the features. Finally, all the dense feature embeddings are concatenated into an embedding matrix that is used by the remaining parts of the model (i.e. feature interaction layers) to predict the probability that a user likes an item (a.k.a. click-through rate prediction). Let be the parameters of feature interaction layers, and be the set of trainable model parameters, the output of the model is given by:
(2) |
where represents the CTR prediction model, is the model prediction. Note that the architecture of feature interaction layers may vary in different CTR prediction models.
The CTR prediction model is trained to optimize the following objective:
(3) |
where is the size of the training dataset and is the model loss function (binary cross-entropy in usual) along with a weight decay regularization term.
Method
In this section, we first formalize the problem of neural feature selection (NFS) and discuss existing solutions that are adapted from gradient-based NAS and network pruning. Then we introduce our Ensemble Gating algorithm that aims to offer effective uncertainty-driven exploration to overcome the gating overfitting problem.
Neural Feature Selection
The goal of neural feature selection is to select features from the feature fields. In this paper, we formulate the NFS problem as a differentiable network pruning problem, thereby connecting it to gradient-based network pruning methods (Xia, Zigeng, and Sanguthevar 2019; Ye et al. 2020) and the problem of gradient-based NAS (Han, Ligeng, and Song 2019). In this formulation, an “over-parameterized” pretrained network is provided, and our task is to remove less informative or redundant model inputs. To automate the feature selection process, a feature gating layer with trainable weights is inserted between the embedding layers and the feature interaction layers (Fig. 1). Specifically, the feature gating layer contains real-valued gating parameters , which matches the number of input feature fields. Then for each gating parameter , a differentiable binarize function converts it into binary gate that determines whether to prune or keep feature embedding :
(4) |
Accordingly, the input to the feature interaction layers is replaced with the masked feature embeddings :
(5) |
Note that the choice of the binarize function is flexible as long as the function can (coarsely) map any real value into and the backward gradient can be accurately estimated. Two commonly used binarize functions include softmax function (with properly picked temperature) and step function with straight through estimator (STE) (Bengio, Léonard, and Courville 2013; Hubara et al. 2016). In this paper, we are most interested in improving the exploration capacity of the STE-based solution, in which the binary gating decision is given by:
(6) |
During the backward propagation, the gradient is computed as if the function were an identity function. For simplicity, we assume the Gating method is implemented with step function and STE in the remainder of this paper.
To optimize the feature selection, the gating parameters are trained to minimize the following loss:
(7) |
where is the dataset used for gating parameters training, is the sparse regularization term that controls the degree of sparsity, and is the weight of sparse regularization. is typically defined with the difference between the target number of selected features and the number of open gates:
(8) |
Note that we don’t penalize when the number of open gates is lower than the target value, because this usually leads to an increase in training loss and the gradient will push the gating layer to reopen some gate(s) (Xia, Zigeng, and Sanguthevar 2019).
Following previous related works, to yield the best feature selection result, the network parameters and the gating parameters are updated iteratively: when training network parameters, all binary gates are fixed and is updated according to Eq. (3) with a mini-batch sampled from ; when training gating parameters, the network parameters are frozen and is updated according to Eq. (7) with a mini-batch sampled from . Note that it’s not necessary that and are different subsets of data. In fact, our experiment results suggest that letting differ from doesn’t yield better results with Gating method (details can be found in the Experiments section). Considering making can be more computationally efficient (because we only need to do one single backward pass at each update step when network parameters and gating parameters share the same input data), we use as the default setting in this paper.
Neural Feature Selection via Ensemble of Gating Layers
One limitation of Gating method is that the gating decision is optimized in a fully exploitative manner, in which the only objective is to minimize the training loss by iteratively updating network parameters and gating parameters. In the Introduction section, we already show that this joint optimization doesn’t provide sufficient gating exploration and can easily get trapped into sub-optimal local minima. Moreover, although the differentiable binarize function relaxes the NP-hard discrete feature selection problem to a continuous optimization problem, the problem itself is still an extremely complex search problem in essence. Hence, a better exploration strategy can definitely benefit the process of training gating parameters.
To mitigate the gating overfitting issue and achieve more effective gating exploration, we present a simple but effective ensemble learning method called Ensemble Gating. Rather than having only one group of binary gates as in Gating, we maintain groups of gates . All the groups share the same underlying network parameters , and for each mini-batch of training data , one group of gates is randomly selected to perform parameters update:
(9) |
This training process is essentially the process of generating bootstrapped subsets of gating training data that is sampled with replacement from the entire dataset. Hence, the objective can also be written as:
(10) |
where represents the bootstrapped training samples for the group of binary gates.
Uncertainty-Driven Exploration Intuitively, the exploration in Ensemble Gating is driven by the inter-group disagreements. For most important features, there can be a significant increase in training loss if they are dropped by the gates. So all the groups can soon reach a consensus on keeping those features. On the contrary, for other less informative features, the gating layer might need to take more exploratory actions (by varying gating decisions and comparing the consequences) to determine their influence on the model performance. Randomly selecting one group to perform parameters update is actually simulating this exploration behavior. In fact, ensemble of neural networks (NNs) has been widely used as a means to estimate predictive uncertainty (Lakshminarayanan, Pritzel, and
Blundell 2016). In Reinforcement Learning research, similar ensemble methods are also used to encourage more meaningful temporally-extended (deep) exploration (Osband et al. 2016; Pathak, Gandhi, and Gupta 2019). In the context of NFS, the agreement/disagreement among the ensemble of binary gates can be interpreted as the gating layer’s predictive uncertainty over feature importance. Only feature embeddings that are considered important with high confidence get fine-tuned constantly. While for features whose importance is yet to be determined, the gating layer will continuously explore different gating decisions. As a consequence, those feature embeddings get fine-tuned more equally and less frequently, thereby effectively preventing the gating overfitting issue.
Input: Training set , iter
Parameter: Pretrained model with parameters , ensemble size , initialized gating parameters
Output: Final gate states
Weight Initialization We increase the inter-group diversity by randomly initializing the gating parameters according to the following uniform distribution as a means to more evenly spread the gating parameters in the search space and to lower the chance to get trapped in any unexpected local minima:
(11) |
Here, controls the percentage of open gates at the beginning (default value: 0.8), and determines the magnitude of gate weights (default value: 0.01). Note that Gating and other related NAS methods usually initialize gating parameters to a constant value to make sure every corresponding feature embedding gets considered and fine-tuned at the beginning. However, this is not a concern in Ensemble Gating because for any feature embedding , the probability that are all initialized to negative values is very low.
Ensemble Aggregation To combine the decisions of different groups, we propose three candidate result aggregation methods:
-
•
Majority Voting (Voting): we sum up the binarized gating decision from all the groups and take the features that receive the most votes as the final output.
-
•
Averaging (Avg): we compute the average of gating parameters and the output is given by the binarized .
-
•
Minimal Retraining Loss (Min): we retrain CTR prediction models from scratch with a few training samples (one epoch of training data in our case). Input features are selected according to the results of each group. The final decision is given by the output of the group that yields minimal (average) retraining loss. Note that this aggregation method is more costly than the other two due to the additional model training.
Algorithm 1 presents the overall training process of Ensemble Gating. As a summary, we would like to highlight some advantages of Ensemble Gating:
-
•
It provides meaningful uncertainty-driven exploration in order to search for the best subset of features.
-
•
Considering the number of input features is much smaller than the number of parameters in a deep CTR prediction model, creating multiple gating layers doesn’t significantly increase memory usage. Hence, Ensemble Gating is a scalable method. Also, its parallelizable nature makes it possible to simultaneously update multiple groups of binary gates, and thereby making it well suited for any distributed learning framework.
-
•
Random weight initialization and random group selection (randomly selecting one group of gates at each update step) work as additional intrinsic randomization to lower the chance of getting stuck into any local minima.
-
•
Ensemble Gating doesn’t make any assumption on the underlying gradient-based feature selection method. Although in this paper we use it to improve the Gating method, it can also apply to other methods with minimum adaptation.
Experiments
Dataset | Model | All Features | Random | Gating | Ensemble Avg | Ensemble Voting | Ensemble Min |
---|---|---|---|---|---|---|---|
Criteo | DCN | 81.37 | 79.857 (0.0052) | 80.725 (0.0023) | 81.014 (0.0001) | 81.006 (0.0003) | 81.028 (0.0001) |
AutoInt | 81.28 | 79.273 (0.0061) | 80.673 (0.0013) | 80.967 (0.0003) | 80.983 (0.0003) | 80.938 (0.0005) | |
DeepFM | 80.43 | 78.737 (0.0031) | 80.22 (0.0011) | 80.336 (0.0006) | 80.29 (0.0005) | 80.336 (0.0005) | |
Avazu | DCN | 78.72 | 73.527 (0.0195) | 77.612 (0.0052) | 78.178 (0.001) | 78.184 (0.001) | 78.184 (0.0007) |
AutoInt | 78.81 | 72.543 (0.0039) | 77.882 (0.0034) | 78.17 (0.0007) | 78.17 (0.0007) | 78.162 (0.0009) | |
DeepFM | 78.0 | 74.307 (0.0126) | 75.59 (0.0286) | 77.592 (0.001) | 77.404 (0.0048)) | 77.618 (0.0008) |
In this section, we present the experiments that evaluate the effectiveness of Ensemble Gating. Firstly, we intend to investigate if our ensemble method can consistently improve Gating. To this end, we compare Ensemble Gating against Gating on two public benchmark datasets (Avazu111https://www.kaggle.com/c/avazu-ctr-prediction and Criteo222https://www.kaggle.com/c/criteo-display-ad-challenge) with three different underlying deep CTR prediction models, namely DeepFM (Guo et al. 2017), AutoInt (Song et al. 2019), and DCN (Wang et al. 2017). Note that depending on the ensemble aggregation method used, there can three different implementations of Ensemble Gating, which are referred to as Ensemble Avg, Ensemble Voting, and Ensemble Min, respectively. Then, to get more insights into Ensemble Gating, we perform a set of complementary experiments, such as varying the numbers of gating layers (ensemble size), and initializing gating parameters in different ways.
Datasets and Tasks
To comply with the policies of most companies, we refrain from presenting the empirical results obtained on any industrial large-scale dataset. Instead, we chose to use two largest open CTR prediction datasets, namely Avazu and Criteo. In Avazu, there are 22 feature fields and the task is to select 6 features that have the highest explanatory power. In Criteo, there are totally 39 feature fields and the task is to select a subset of 19 features.
We randomly split the dataset into 80% and 20% for training and testing. The testing dataset is further randomly split in half for validation and testing in our complementary experiments where network parameter update and gating parameter update are performed on different subsets of data. The evaluation metric we adopt is the AUC (Area Under the ROC Curve) achieved on the testing set by the model that is retrained with the selected features.
Training Setup
Both Ensemble Gating and the baselines follow the same training process. Firstly, we pretrain a CTR prediction model with all input features for 10 epochs. We apply the Adam optimizer (Kingma and Ba 2015) with a learning rate of 0.001 for network parameters . The mini-batch size is set to 2048 and the weight of weight decay regularizer is set to . The weights of pretrained model are saved for later steps.
After obtaining the pretrained model, the gating layer(s) will be inserted between the embedding layers and the interaction layers. Then we iteratively updated the network parameters and the gating parameters for 8 epochs. We use SGD with a learning rate of 0.001 to optimize the gating parameters.
Following the Lottery Ticket Hypothesis (Frankle and Carbin 2018), our last step is to retrain the model from scratch with selected features on the training set (for 10 epochs) and test it on the testing set. The final gating decision of Ensemble Gating is calculated according to the particular aggregation method. In our preliminary experiments, we observed that in both Gating and Ensemble Gating, the number of open gates at the end of training is occasionally one or two smaller than the target value . The most likely cause of this phenomenon is the choice of the hyperparameter (weight of sparse regularizer). Considering that finding the “optimal” value of can be very expensive in practice, we adopt a simple workaround, which adds unselected features with the largest gate weights (average gate weights in ensemble method) to the set of selected features. This process resembles adjusting the gating threshold that is set to zero in the binarized function (Eq. 6). Our preliminary results suggest that this post-processing doesn’t have any observable negative impact on the final outcome.
Experiment Results
We ran each feature selection method 5 times and recorded the AUC score achieved on the testing set. Tab. 1 and Fig. 4 in Appendix A present the results of Ensemble Gating and the baselines. For reference, we also include the results of the random baseline (randomly selecting features) and the full-features baseline (using all the features in retraining). The results show that both Gating and Ensemble Gating outperform the random baseline by a large margin, and Ensemble Gating is able to further improve Gating by selecting a better subset of features that results in a significantly higher average test AUC. Moreover, Ensemble Gating also has a noticeably smaller standard deviation of AUC scores, which suggests that Ensemble Gating’s uncertain-driven exploration enables it to more consistently find a set of informative features regardless of all the randomization (e.g. dataset shuffling) and its initial position in the search space.
One might argue that the mean and standard deviation do not tell the whole story, because when there are adequate computational resources, people can run the searching algorithm for multiple times and pick the best one as the final output. To address this concern, we rank the results of Gating and Ensemble Avg according to their AUC scores, and count how many of the top three results were obtained with the ensemble method. We refer to this value as Top-3 score. On Criteo, our Top-3 scores are 3 (DCN), 3 (AutoInt), and 2 (DeepFM); on Avazu, the scores are 3 (DCN), 3 (AutoInt), and 3 (DeepFM). This confirms that Ensemble Gating exhibits better exploratory capacity and manages to constantly find a better subset of features.
Convergence Discussion As pointed out by many previous works (Xia, Zigeng, and Sanguthevar 2019; Gordon et al. 2018), there is no strong guarantee of convergence in Gating, but empirical results suggest that Gating usually converges to a fairly well state within a few epochs in practice. Hence, in this section, we intend to examine whether this fast “convergence” property holds in our ensemble method. To this end, we keep track of the average number of open gates and inter-group differences in the searching/training process. The inter-group difference is measured by counting the number of features on which two groups disagree. Here, we take the results obtained on Avazu with AutoInt as example (Fig. 3). The plots show that the number of open gates is quickly optimized to the target value and the inter-group difference drops to a reasonably small value within two epochs. Note that theoretically the inter-group difference is not necessarily zero at convergence because of the non-convex nature of the feature selection problem and the fact that there might exist multiple optimal solutions.
Additional Experiments
In this section, we will conduct a set of experiments with the aim of getting more insights into Ensemble Gating. We assume the Average aggregation method is used in all the complementary experiments.
What is a good value for the size of ensemble? The answer to this question can be found in Fig. 4 in Appendix A, in which we vary the number of binary gate groups. It is clear that the performance of 5 groups and 10 groups is better than that of 2 groups, and 2 groups outperforms Gating in only a few cases. Hence, we can conclude that an ensemble of size 5 is sufficient to capture the benefits of the ensemble method.
Is our uncertainty-driven exploration a better strategy compared to other methods? To answer this question, we compare Ensemble Gating to an approach that uses another popular binarize function Gumbel-Softmax (Jang, Gu, and Poole 2016). Since each binary gate only has two states (open or closed), Gumbel-Softmax function in this case is essentially a Gumbel-Sigmoid function as in (Tsai et al. 2018):
(12) |
where is a Gumbel noise, is a uniform random variable in the range of , and is a small enough temperature parameter. Gumbel-Sigmoid function becomes closer to a true binarize function when approaches zero. Similar to (Xie et al. 2018), we gradually decrease the value of from to throughout the searching/training process. We also experimented with other temperature annealing configurations, but results remained similar.
Gumbel-Sigmoid function can potentially improve gating exploration due to the existence of the Gumbel noise. It resembles -greedy exploration in Reinforcement Learning in a sense that both of them add random exploration incentive without considering state context. On the contrary, Ensemble Gating’s exploration strategy is based on predictive uncertainty, which is highly state dependent. Tab. 2 compares the performance of Ensemble Gating and Gumbel-Sigmoid, and it shows that uncertainty-driven exploration works better than simply adding random noise into the system.
Dataset | Model | Gumbel-Sigmoid | Gating | Ensemble Avg |
---|---|---|---|---|
Criteo | DCN | 80.427 (0.0026) | 80.725 (0.0023) | 81.014 (0.0001) |
AutoInt | 80.472 (0.0011) | 80.673 (0.0013) | 80.967 (0.0003) | |
DeepFM | 79.813 (0.0017) | 80.22 (0.0011) | 80.336 (0.0006) | |
Avazu | DCN | 77.254 (0.0049) | 77.612 (0.0052) | 78.178 (0.001) |
AutoInt | 77.325 (0.0075) | 77.882 (0.0034) | 78.17 (0.0007) | |
DeepFM | 76.636 (0.0065) | 75.59 (0.0286) | 77.592 (0.001) |
Is random weight initialization employed by Ensemble Gating a key factor that distinguishes Ensemble Gating from Gating? One difference between Gating and Ensemble Gating is the way they initialize the gate weights. Here we perform additional experiments to examine the effect of different initialization methods on the final outcome. We refer to Ensemble Gating with random initialization as Ensemble Random, and our ensemble method with constant initialization as Ensemble Constant. We follow the same naming convention for Gating. The results are presented in Tab. 3. We can observe that the performance of Ensemble Random and Ensemble Constant are very close while Ensemble Random still maintains a slight edge over other approaches. Not surprisingly, random weight initialization degrades the performance of Gating. This can be explained by the gating overfitting problem too: the network parameters are fine-tuned to prefer features with gates that are initialized to be open. On the contrary, as mentioned in the previous section, having a relatively small number of gates closed initially is not a concern in ensemble method. Hence, Ensemble Gating can benefit from the inter-group diversity that arises from random initialization.
Model | Gating Constant | Gating Random | Ensemble Random | Ensemble Constant | |
---|---|---|---|---|---|
Criteo | DCN | 80.725 (0.0023) | 80.516 (0.0042) | 81.014 (0.0001) | 80.96 (0.0005) |
AutoInt | 80.673 (0.0013) | 80.74 (0.0018) | 80.967 (0.0003) | 80.93 (0.0004) | |
DeepFM | 80.22 (0.0011) | 80.133 (0.0013) | 80.336 (0.0006) | 80.273 (0.0007) | |
Avazu | DCN | 77.612 (0.0052) | 77.504 (0.0033) | 78.178 (0.001) | 78.125 (0.001) |
AutoInt | 77.882 (0.0034) | 77.122 (0.0125) | 78.17 (0.0007) | 78.195 (0.0001) | |
DeepFM | 75.59 (0.0286) | 74.39 (0.0199) | 77.592 (0.001) | 77.677 (0.0005) |
How important is it to update the binary gates on a separate validation set? Some related works suggest training gating parameters on a separate validation dataset in order to enhance the generalization of the gating result (Zhao et al. 2021; Xia, Zigeng, and Sanguthevar 2019; Han, Ligeng, and Song 2019). However, in our experiments, the performance of Gating doesn’t get improved when is different from (see Appendix A for all the results). One possible reason for this is the validation set in Avazu and Criteo might be not large enough to capture the distribution of all the data. Nevertheless, Ensemble Gating is able to outperform Gating under both settings (i.e. and ).
Can we freeze network parameters to prevent the gating overfitting issue? Most gradient-based NAS and network pruning methods follow the multi-step training/searching framework, which iteratively updates the sparsity structure and fine-tune the original model weights. We show in previous section that the major drawback of this multi-step training framework is the model weights can end up “overfitting” to certain gating state. Hence, one natural question is whether we can bypass the model fine-tuning to avoid the gating overfitting issue. Our experimental results confirm that fine-tuning the network parameters is an indispensable step to find a good subset of features (see Appendix A for all the results). Hence, the right way to improve differentiable neural feature selection methods should be looking for better exploration strategy rather than bypassing the fine-tuning steps.
Conclusion & Future Work
In this work, we presented a novel ensemble method that offers a more efficient exploration strategy to find the optimal subset of features. We show that our proposed method, Ensemble Gating, consistently improves Gating method on two public datasets with three different underlying CTR prediction models. The proposed method is computationally efficient and can be easily applied to a variety of CTR prediction models.
Future work may include studying whether our ensemble method can benefit related research problems, such as differentiable network pruning and neural input search. Insufficient exploration can be detrimental in any search problem, and gating overfitting may exist as long as similar gating formulations are employed. Moreover, we note that we have used our ensemble method to improve Gating method implemented with step function and STE. Future work can experiment with other binarize functions.
References
- Bengio, Léonard, and Courville (2013) Bengio, Y.; Léonard, N.; and Courville, A. C. 2013. Estimating or Propagating Gradients Through Stochastic Neurons for Conditional Computation. ArXiv, abs/1308.3432.
- Chandrashekar and Sahin (2014) Chandrashekar, G.; and Sahin, F. 2014. A survey on feature selection methods. Computers & Electrical Engineering, 40(1): 16–28.
- Cheng, Shen, and Huang (2020) Cheng, W.; Shen, Y.; and Huang, L. 2020. Differentiable neural input search for recommender systems. arXiv preprint arXiv:2006.04466.
- Frankle and Carbin (2018) Frankle, J.; and Carbin, M. 2018. The lottery ticket hypothesis: Finding sparse, trainable neural networks. arXiv preprint arXiv:1803.03635.
- Gao et al. (2020) Gao, S.; Huang, F.; Pei, J.; and Huang, H. 2020. Discrete model compression with resource constraint for deep neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 1899–1908.
- Ginart et al. (2019) Ginart, A.; Naumov, M.; Mudigere, D.; Yang, J.; and Zou, J. 2019. Mixed dimension embeddings with application to memory-efficient recommendation systems. arXiv preprint arXiv:1909.11810.
- Gordon et al. (2018) Gordon, A.; Eban, E.; Nachum, O.; Chen, B.; Wu, H.; Yang, T.-J.; and Choi, E. 2018. Morphnet: Fast & simple resource-constrained structure learning of deep networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, 1586–1595.
- Guo et al. (2017) Guo, H.; Tang, R.; Ye, Y.; Li, Z.; and He, X. 2017. DeepFM: a factorization-machine based neural network for CTR prediction. In IJCAI’17 Proceedings of the 26th International Joint Conference on Artificial Intelligence, 1725–1731.
- Han, Ligeng, and Song (2019) Han, C.; Ligeng, Z.; and Song, H. 2019. ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware. In International Conference on Learning Representations.
- Han et al. (2015) Han, S.; Pool, J.; Tran, J.; and Dally, W. J. 2015. Learning both weights and connections for efficient neural networks. arXiv preprint arXiv:1506.02626.
- Hubara et al. (2016) Hubara, I.; Courbariaux, M.; Soudry, D.; El-Yaniv, R.; and Bengio, Y. 2016. Binarized neural networks. Advances in neural information processing systems, 29.
- Jang, Gu, and Poole (2016) Jang, E.; Gu, S.; and Poole, B. 2016. Categorical Reparameterization with Gumbel-Softmax. In ICLR (Poster).
- Joglekar et al. (2019) Joglekar, M. R.; Li, C.; Adams, J. K.; Khaitan, P.; and Le, Q. V. 2019. Neural Input Search for Large Scale Recommendation Models. CoRR abs/1907.04471 (2019). arXiv preprint arXiv:1907.04471.
- Kingma and Ba (2015) Kingma, D. P.; and Ba, J. L. 2015. Adam: A Method for Stochastic Optimization. In ICLR 2015 : International Conference on Learning Representations 2015.
- Koenigstein and Paquet (2013) Koenigstein, N.; and Paquet, U. 2013. Xbox movies recommendations: Variational Bayes matrix factorization with embedded feature selection. In Proceedings of the 7th ACM Conference on Recommender Systems, 129–136.
- Lakshminarayanan, Pritzel, and Blundell (2016) Lakshminarayanan, B.; Pritzel, A.; and Blundell, C. 2016. Simple and scalable predictive uncertainty estimation using deep ensembles. arXiv preprint arXiv:1612.01474.
- Liu, Simonyan, and Yang (2018) Liu, H.; Simonyan, K.; and Yang, Y. 2018. Darts: Differentiable architecture search. arXiv preprint arXiv:1806.09055.
- Liu et al. (2021) Liu, S.; Gao, C.; Chen, Y.; Jin, D.; and Li, Y. 2021. LEARNABLE EMBEDDING SIZES FOR RECOMMENDER SYSTEMS. In ICLR 2021: The Ninth International Conference on Learning Representations.
- Maddison, Mnih, and Teh (2016) Maddison, C. J.; Mnih, A.; and Teh, Y. W. 2016. The concrete distribution: A continuous relaxation of discrete random variables. arXiv preprint arXiv:1611.00712.
- Osband et al. (2016) Osband, I.; Blundell, C.; Pritzel, A.; and Van Roy, B. 2016. Deep exploration via bootstrapped DQN. Advances in neural information processing systems, 29: 4026–4034.
- Parmezan et al. (2021) Parmezan, A. R. S.; Lee, H. D.; Spolaôr, N.; and Wu, F. C. 2021. Automatic recommendation of feature selection algorithms based on dataset characteristics. Expert Systems with Applications, 185: 115589.
- Pathak, Gandhi, and Gupta (2019) Pathak, D.; Gandhi, D.; and Gupta, A. 2019. Self-supervised exploration via disagreement. In International conference on machine learning, 5062–5071. PMLR.
- Ronen et al. (2013) Ronen, R.; Koenigstein, N.; Ziklik, E.; and Nice, N. 2013. Selecting content-based features for collaborative filtering recommenders. In Proceedings of the 7th ACM conference on Recommender systems, 407–410.
- Sheth and Fusi (2020) Sheth, R.; and Fusi, N. 2020. Differentiable Feature Selection by Discrete Relaxation. In International Conference on Artificial Intelligence and Statistics, 1564–1572. PMLR.
- Song et al. (2019) Song, W.; Shi, C.; Xiao, Z.; Duan, Z.; Xu, Y.; Zhang, M.; and Tang, J. 2019. Autoint: Automatic feature interaction learning via self-attentive neural networks. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, 1161–1170.
- Srinivas, Subramanya, and Venkatesh Babu (2017) Srinivas, S.; Subramanya, A.; and Venkatesh Babu, R. 2017. Training sparse neural networks. In Proceedings of the IEEE conference on computer vision and pattern recognition workshops, 138–145.
- Tsai et al. (2018) Tsai, Y.-H.; Liu, M.-Y.; Sun, D.; Yang, M.-H.; and Kautz, J. 2018. Learning binary residual representations for domain-specific video streaming. In Thirty-second AAAI conference on artificial intelligence.
- Wang et al. (2017) Wang, R.; Fu, B.; Fu, G.; and Wang, M. 2017. Deep & cross network for ad click predictions. In Proceedings of the ADKDD’17, 1–7.
- Xia, Zigeng, and Sanguthevar (2019) Xia, X.; Zigeng, W.; and Sanguthevar, R. 2019. AutoPrune: Automatic Network Pruning by Regularizing Auxiliary Parameters. In Advances in Neural Information Processing Systems 32 (NeurIPS 2019).
- Xie et al. (2018) Xie, S.; Zheng, H.; Liu, C.; and Lin, L. 2018. SNAS: stochastic neural architecture search. arXiv preprint arXiv:1812.09926.
- Yamada et al. (2020) Yamada, Y.; Lindenbaum, O.; Negahban, S.; and Kluger, Y. 2020. Feature selection using stochastic gates. In International Conference on Machine Learning, 10648–10659. PMLR.
- Ye et al. (2020) Ye, M.; Choudhary, D.; Yu, J.; Wen, E.; Chen, Z.; Yang, J.; Park, J.; Liu, Q.; and Kejariwal, A. 2020. Adaptive Dense-to-Sparse Paradigm for Pruning Online Recommendation System with Non-Stationary Data. arXiv preprint arXiv:2010.08655.
- Zhao et al. (2021) Zhao, X.; Liu, H.; Liu, H.; Tang, J.; Guo, W.; Shi, J.; Wang, S.; Gao, H.; and Long, B. 2021. AutoDim: Field-aware Embedding Dimension Searchin Recommender Systems. In Proceedings of the Web Conference 2021, 3015–3022.
Appendix A Appendix
A. Results of All Experiments
In this section, we present all the experimental results, some of which were not included in the paper due to space constraint.
Testing AUC scores of Ensemble Gating and Gating
The results are visualized in Fig. 4.






Comparison between Ensemble Gating and Gumbel-Sigmoid
We intend to verify whether the uncertainty-driven exploration can be more efficient than other exploration strategy. The results are visualized in Fig. 5 and Fig. 6.






Comparison between Different Weights Initialization Methods
We used two different methods, i.e. random initialization and constant initialization, to initialize the binary gates. The results can be found in Fig. 7 and Fig. 8.






Comparison between Searching on Validation Set and Searching on Training Set
We intend to investigate the importance of training gating parameters on a separate validation set. We denote the methods that use separate validation set as Gating Val and Ensemble Val. The results can be found in Fig. 9 and Fig. 10.





