888.470760_415140.lt. [POPULAR]
Recommender systems often struggle to balance memorization (learning frequent, specific co-occurrences of items/features) and generalization (recommending items that haven't explicitly appeared together in the training data) [1606.07792].
A wide linear model is used, which excels at memorizing sparse feature interactions (e.g., user clicked 'item A' and user is from 'location B') [1606.07792].
The paper proposes training both components simultaneously rather than separately. This allows the model to optimize for both accuracy (via the wide component) and serendipity/novelty (via the deep component) [1606.07792]. Key Results & Impact 888.470760_415140.lt.
The query likely refers to the seminal 2016 paper published by researchers at Google [1606.07792]. This paper introduced a model that combines the strengths of linear models (memorization) and deep neural networks (generalization) to improve recommendation quality. Core Concepts of the "Wide & Deep" Paper
Online experiments showed that "Wide & Deep" significantly increased app acquisitions compared to models that used either approach alone [1606.07792]. This allows the model to optimize for both
Discuss the used in the model (e.g., user, context, item features).
This architecture has since become a standard baseline for many recommendation tasks in industry, including those described in studies on YouTube recommendations [1606.07792]. If you'd like, I can: Core Concepts of the "Wide & Deep" Paper
A deep feed-forward neural network is used, which generalizes better to unseen feature combinations by learning low-dimensional dense embeddings for sparse features [1606.07792].