RecSys 2021 - Papers and Talks to Chew on

[ recsys deeplearning production survey ] · 5 min read

RecSys 2021 happened this week (27 Sept - 1 Oct). Here are some papers I found interesting.

Negative Interactions for Improved Collaborative-Filtering: Don’t go Deeper, go Higher was motivated by the finding that modeling higher-order interactions helps with recommendation accuracy. They shared a simple extension of adding higher-order interactions to a linear model without a hidden layer (Embarrassingly Shallow AutoEncoders aka EASE^R). EASE^R learns pairwise relationships between each item i (input) and item j (output) of the autoencoder.

To add higher-order interactions as input, two items are now considered as input (i and k) to predict item j in the output. (They also tried even higher-order interactions but didn’t see any improvements.) This simple extension on a simple model was competitive with several SOTA deep learning models on datasets such as MovieLens-20M, Netflix, and MSD.

Learning higher-order interactions with original pairwise interactions

Learning higher-order interactions with original pairwise interactions (source)

In addition, the paper showed that less active users benefited more from the higher-order model relative to EASE^R. They hypothesized that, because triplet-relations (i, k, j) are more prevalent among highly-active users, the pairwise relationships are freed up to better adapt to less-active users. Loved the simplicity of this idea and implementation.

Reenvisioning the comparison between Neural Collaborative Filtering and Matrix Factorization revisits the comparison between matrix factorization (MF) and neural collaborative filtering (NCF) again.

To recap, MF learns a latent representation of items and users and combines these representations to compute a preference score between each user and item (e.g., dot product). In comparison, NCF uses multilayer perceptrons (or other deep learning layers) to learn scores between each user and item.

The current paper reproduces the results from a RecSys 2020 paper that compared MF and NCF, and extends it by including other accuracy metrics, as well as metrics for diversity and novelty. It showed that MF outperforms NCF in performance (nDCG and Hit Rate), including in the long tail, though NCF provides more diversity and novelty. The paper also includes a useful list of various recommendation baselines. Takeaway: Don’t throw your MF techniques out yet.

You Do Not Need a Bigger Boat: Recommendations at Reasonable Scale in a (Mostly) Serverless and Open Stack shares a few principles and a suggested design for deploying recommenders using cloud services and open-source packages.

Suggested design and tech stack for training and serving a recommender system

Suggested design and tech stack for training and serving a recommender system (source)

Principles include focusing on data quality (which leads to bigger gains relative to model improvements), using managed services instead of maintaining and scaling infrastructure, and reduced dependence on distributed computing (e.g., Spark) which can be slow and hard to debug. They also provide an open-sourced implementation of a tech stack that goes from data ingestion (AWS Lambda) to recommendation serving (AWS SageMaker). Batteries (read: open dataset with 30 million rows) included.

Transformers4Rec: Bridging the Gap between NLP and Sequential / Session-Based Recommendation introduces Transformers4Rec, an open-source library built on HuggingFace’s Transformers.

It applies the Transformer architecture (and variants such as GPT-2, BERT, XLNet) to sequential and session-based recommendations. The paper includes results from several experiments, such as using different training regimes (casual language modeling (LM), permutation LM, mask LM) and different ways to integrate side information. Experimenting with session-based recommenders? Try this library out.

RecSysOps: Best Practices for Operating a Large-Scale Recommender System shared a set of best practices for identifying, diagnosing, and resolving issues in large-scale recommendation systems (RecSysOps). Best practices are divided into four categories:

  • Detection: implementing know best practices, monitoring the system end-to-end, understanding why users engage with low ranked items
  • Prediction: predicting items that will have cold-start before launch date (e.g., new shows or movies that are added to catalog)
  • Diagnosis: logging, issue reproducibility, distinguishing between input data issue (e.g., incorrect language) and model issue (e.g., missing values handled incorrectly)
  • Resolution: having a playbook of hotfixes, considering and handling issues (e.g., corrupted data, timeouts) into the system to make it more robust

Semi-Supervised Visual Representation Learning for Fashion Compatibility shares about how they overcame the constraints of limited labeled data for fashion compatibility prediction (e.g., an outfit consisting of dress, jacket, shoes). Their model is a siamese network with a ResNet18 backbone trained on labeled triplets of anchor item, compatible item, and non-compatible item images.

To augment their data, they adopt a semi-supervised learning approach. During training, pseudo positive outfits were generated by replacing compatible items with a nearest neighbor item to get pseudo compatible outfits. Pseudo non-compatible outfits are generated similarly via replacing items.

Creating pseudo-labels (middle) and applying shape and color transformations (right)

Creating pseudo-labels (middle) and applying shape and color transformations (right) (source)

They also observed that compatible items have color and texture similarity, but not shape similarity. Thus, they applied self-supervised consistency regularization where shape and color perturbed images are used as positive and negative labels respectively.

Shared Neural Item Representations for Completely Cold Start Problem shared their findings that using user interaction vectors as input achieves better results in fewer iterations relative to using customer ID as input. (I had this intuition and it’s great to see experiment results on this.) Thus, they use item embeddings to represent users.

Unifying item representations across user and item towers

Unifying item representations across user and item towers (source)

With this approach, two sets of item embeddings are learned—item embeddings to represent the user, and item embeddings to represent items. To simplify and improve learning, they unify the item embeddings by using item embedding learned via the item tower to also represent users. They also include side information when learning item embeddings to handle item cold-start.

What papers did you enjoy? Reach out and let me know!

Share on:

Browse related tags: [ recsys deeplearning production survey ]

Join 2,300+ readers getting updates on data science, ML systems, & career.

    Welcome gift: A 5-day email course on How to be an Effective Data Scientist 🚀