deepctr.models.sequence.din module

Author:
Weichen Shen, weichenswc@163.com
Reference:
[1] Zhou G, Zhu X, Song C, et al. Deep interest network for click-through rate prediction[C]//Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. ACM, 2018: 1059-1068. (https://arxiv.org/pdf/1706.06978.pdf)
deepctr.models.sequence.din.DIN(dnn_feature_columns, history_feature_list, dnn_use_bn=False, dnn_hidden_units=(256, 128, 64), dnn_activation='relu', att_hidden_size=(80, 40), att_activation='dice', att_weight_normalization=False, l2_reg_dnn=0, l2_reg_embedding=1e-06, dnn_dropout=0, seed=1024, task='binary')[source]

Instantiates the Deep Interest Network architecture.

Parameters:
  • dnn_feature_columns – An iterable containing all the features used by deep part of the model.
  • history_feature_list – list,to indicate sequence sparse field
  • dnn_use_bn – bool. Whether use BatchNormalization before activation or not in deep net
  • dnn_hidden_units – list,list of positive integer or empty list, the layer number and units in each layer of deep net
  • dnn_activation – Activation function to use in deep net
  • att_hidden_size – list,list of positive integer , the layer number and units in each layer of attention net
  • att_activation – Activation function to use in attention net
  • att_weight_normalization – bool.Whether normalize the attention score of local activation unit.
  • l2_reg_dnn – float. L2 regularizer strength applied to DNN
  • l2_reg_embedding – float. L2 regularizer strength applied to embedding vector
  • dnn_dropout – float in [0,1), the probability we will drop out a given DNN coordinate.
  • seed – integer ,to use as random seed.
  • task – str, "binary" for binary logloss or "regression" for regression loss
Returns:

A Keras model instance.