https://eugeneyan.com/rss 09月30日
RecSys 2021 论文盘点:深度学习与推荐系统新进展
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

RecSys 2021 大会分享了多项有趣论文,包括通过高阶交互提升协同过滤精度的 EASE^R 模型、重新评估矩阵分解与神经协同过滤的对比研究、云服务与开源工具的推荐系统部署方案、Transformer 在序列推荐中的应用、大规模推荐系统运维最佳实践、时尚兼容性预测的半监督学习方法,以及冷启动问题的统一表示方案。这些研究展示了推荐系统在深度学习、数据质量、可扩展性和跨领域应用方面的最新进展。

🔍 EASE^R 模型通过引入高阶交互(如三元组关系 i, k, j)扩展线性模型,在 MovieLens-20M 等数据集上与深度学习模型性能相当,且对低活跃度用户更有效,因活跃用户更密集的三元组关系释放了更多对低活跃用户的适配资源。

📈 矩阵分解(MF)在 nDCG 和 Hit Rate 等指标上优于神经协同过滤(NCF),尤其在长尾推荐中表现更优,但 NCF 提供更高推荐多样性和新颖性,提示 MF 技术仍有价值。

⚙️ 推荐系统可合理部署于云服务(如 AWS Lambda for data ingestion, SageMaker for serving),优先提升数据质量而非模型复杂度,减少对 Spark 等分布式计算的依赖,并开源了完整技术栈实现。

🗣️ Transformers4Rec 将 Transformer 架构(如 GPT-2)应用于序列/会话推荐,通过语言模型训练(LM)和侧信息整合提升效果,为会话推荐提供新工具。

🛠️ RecSysOps 提出四大运维最佳实践:检测(全链路监控)、预测(冷启动物品预测)、诊断(日志与问题复现)和修复(热修复与系统健壮性设计),覆盖系统全生命周期。

👗 时尚兼容性预测采用半监督学习,通过 Siamese 网络(ResNet18 backbone)学习兼容/不兼容三元组图像表示,并利用自监督一致性正则化(形状扰动为负样本)缓解标注稀缺问题。

🔥 冷启动问题通过统一用户和物品嵌入表示解决,使用物品塔学习的嵌入同时代表用户,结合侧信息处理物品冷启动,且用户交互向量输入比用户 ID 收敛更快。

RecSys 2021 happened this week (27 Sept - 1 Oct). Here are some papers I found interesting.

Negative Interactions for Improved Collaborative-Filtering: Don’t go Deeper, go Higher was motivated by the finding that modeling higher-order interactions helps with recommendation accuracy. They shared a simple extension of adding higher-order interactions to a linear model without a hidden layer (Embarrassingly Shallow AutoEncoders aka EASE^R). EASE^R learns pairwise relationships between each item i (input) and item j (output) of the autoencoder.

To add higher-order interactions as input, two items are now considered as input (i and k) to predict item j in the output. (They also tried even higher-order interactions but didn’t see any improvements.) This simple extension on a simple model was competitive with several SOTA deep learning models on datasets such as MovieLens-20M, Netflix, and MSD.

Learning higher-order interactions with original pairwise interactions (source)

In addition, the paper showed that less active users benefited more from the higher-order model relative to EASE^R. They hypothesized that, because triplet-relations (i, k, j) are more prevalent among highly-active users, the pairwise relationships are freed up to better adapt to less-active users. Loved the simplicity of this idea and implementation.

Reenvisioning the comparison between Neural Collaborative Filtering and Matrix Factorization revisits the comparison between matrix factorization (MF) and neural collaborative filtering (NCF) again.

To recap, MF learns a latent representation of items and users and combines these representations to compute a preference score between each user and item (e.g., dot product). In comparison, NCF uses multilayer perceptrons (or other deep learning layers) to learn scores between each user and item.

The current paper reproduces the results from a RecSys 2020 paper that compared MF and NCF, and extends it by including other accuracy metrics, as well as metrics for diversity and novelty. It showed that MF outperforms NCF in performance (nDCG and Hit Rate), including in the long tail, though NCF provides more diversity and novelty. The paper also includes a useful list of various recommendation baselines. Takeaway: Don’t throw your MF techniques out yet.

You Do Not Need a Bigger Boat: Recommendations at Reasonable Scale in a (Mostly) Serverless and Open Stack shares a few principles and a suggested design for deploying recommenders using cloud services and open-source packagehttps://eugeneyan.com/assets/ydnabb.webp="/assets/ydnabb.webp" loading="lazy" title="Suggested design and tech stack for training and serving a recommender system" alt="Suggested design and tech stack for training and serving a recommender system">

Suggested design and tech stack for training and serving a recommender system (source)

Principles include focusing on data quality (which leads to bigger gains relative to model improvements), using managed services instead of maintaining and scaling infrastructure, and reduced dependence on distributed computing (e.g., Spark) which can be slow and hard to debug. They also provide an open-sourced implementation of a tech stack that goes from data ingestion (AWS Lambda) to recommendation serving (AWS SageMaker). Batteries (read: open dataset with 30 million rows) included.

Transformers4Rec: Bridging the Gap between NLP and Sequential / Session-Based Recommendation introduces Transformers4Rec, an open-source library built on HuggingFace’s Transformers.

It applies the Transformer architecture (and variants such as GPT-2, BERT, XLNet) to sequential and session-based recommendations. The paper includes results from several experiments, such as using different training regimes (casual language modeling (LM), permutation LM, mask LM) and different ways to integrate side information. Experimenting with session-based recommenders? Try this library out.

RecSysOps: Best Practices for Operating a Large-Scale Recommender System shared a set of best practices for identifying, diagnosing, and resolving issues in large-scale recommendation systems (RecSysOps). Best practices are divided into four categories:

    Detection: implementing know best practices, monitoring the system end-to-end, understanding why users engage with low ranked items Prediction: predicting items that will have cold-start before launch date (e.g., new shows or movies that are added to catalog) Diagnosis: logging, issue reproducibility, distinguishing between input data issue (e.g., incorrect language) and model issue (e.g., missing values handled incorrectly) Resolution: having a playbook of hotfixes, considering and handling issues (e.g., corrupted data, timeouts) into the system to make it more robust

Semi-Supervised Visual Representation Learning for Fashion Compatibility shares about how they overcame the constraints of limited labeled data for fashion compatibility prediction (e.g., an outfit consisting of dress, jacket, shoes). Their model is a siamese network with a ResNet18 backbone trained on labeled triplets of anchor item, compatible item, and non-compatible item images.

To augment their data, they adopt a semi-supervised learning approach. During training, pseudo positive outfits were generated by replacing compatible items with a nearest neighbor item to get pseudo compatible outfits. Pseudo non-compatible outfits are generated similarhttps://eugeneyan.com/assets/fashion-compatibility.webpmg src="/assets/fashion-compatibility.webp" loading="lazy" title="Creating pseudo-labels (middle) and applying shape and color transformations (right)" alt="Creating pseudo-labels (middle) and applying shape and color transformations (right)">

Creating pseudo-labels (middle) and applying shape and color transformations (right) (source)

They also observed that compatible items have color and texture similarity, but not shape similarity. Thus, they applied self-supervised consistency regularization where shape and color perturbed images are used as positive and negative labels respectively.

Shared Neural Item Representations for Completely Cold Start Problem shared their findings that using user interaction vectors as input achieves better results in fewer iterations relative to using customer ID as input. (I had this intuition and it’s great to see experiment results on this.) Thus, https://eugeneyan.comhttps://eugeneyan.com/assets/shared-item-representations.webpers.

Unifying item representations across user and item towers (source)

With this approach, two sets of item embeddings are learned—item embeddings to represent the user, and item embeddings to represent items. To simplify and improve learning, they unify the item embeddings by using item embedding learned via the item tower to also represent users. They also include side information when learning item embeddings to handle item cold-start.

What papers did you enjoy? Reach out and let me know!

If you found this useful, please cite this write-up as:

Yan, Ziyou. (Oct 2021). RecSys 2021 - Papers and Talks to Chew on. eugeneyan.com. https://eugeneyan.com/writing/recsys2021/.

or

@article{yan2021recsys,  title   = {RecSys 2021 - Papers and Talks to Chew on},  author  = {Yan, Ziyou},  journal = {eugeneyan.com},  year    = {2021},  month   = {Oct},  url     = {https://eugeneyan.com/writing/recsys2021/}}
Share on:

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

RecSys 2021 推荐系统 深度学习 协同过滤 Transformer 云服务 运维 半监督学习 冷启动 矩阵分解
相关文章