Awesome-LLM-for-RecSys
Awesome-LLM-for-RecSys 是一个专注于大语言模型(LLM)与推荐系统交叉领域的开源资源库。它系统性地收集并整理了该方向的高质量学术论文与技术资源,旨在帮助从业者快速掌握如何利用大模型提升推荐系统的性能。
当前,传统推荐系统在特征工程、语义理解及冷启动等方面面临瓶颈,而大模型的出现为解决这些问题提供了新范式。Awesome-LLM-for-RecSys 通过独特的分类框架,将论文按大模型在推荐流程中的介入位置(如特征增强、排序优化、生成式推荐等)进行梳理,清晰展示了技术演进脉络。项目不仅包含详尽的论文列表,还配套了被 ACM TOIS 接收的综述文章,并持续更新最新研究成果。
这份资源特别适合人工智能研究人员、算法工程师以及对下一代推荐技术感兴趣的学生使用。无论是希望深入探索 LLM 在推荐场景中应用机理的学者,还是寻求落地解决方案的开发者,都能从中找到极具价值的参考依据和技术灵感,从而高效跟进这一前沿领域的发展动态。
使用场景
某电商平台的算法团队正致力于利用大语言模型(LLM)重构其商品推荐系统,以解决传统模型在理解用户复杂意图和冷启动商品特征提取上的瓶颈。
没有 Awesome-LLM-for-RecSys 时
- 文献检索如大海捞针:团队成员需在 ArXiv 等平台上手动搜索分散的论文,难以系统性掌握 LLM 在特征工程、排序策略等全链路的应用现状。
- 技术选型缺乏依据:面对众多模型(如 LLaMA、ChatGLM)和微调策略(Full Finetuning、LoRA、Prompt Tuning),无法快速对比不同方案在特定场景(如少样本学习)下的优劣。
- 重复造轮子风险高:由于缺乏统一的资源索引,团队可能花费数周复现已被证明效果不佳的方法,或忽略了最新的 SOTA(最先进)成果。
- 跨领域知识融合难:难以发现如何将知识图谱增强、多模态标签生成等前沿技术与现有推荐架构有效结合的具体案例。
使用 Awesome-LLM-for-RecSys 后
- 一站式获取全景视野:直接查阅按推荐系统流水线分类的论文列表,迅速定位到“特征增强”或“生成式推荐”等关键领域的最新研究。
- 精准决策技术路线:参考列表中详细的"LLM 骨干网络”与“微调策略”对比表,快速选定适合电商场景的 LoRA 微调方案,大幅缩短验证周期。
- 紧跟前沿避免滞后:通过定期更新的“最新研究工作列表”,团队能即时采纳如 TOIS 收录的最新综述观点,确保技术方案始终处于行业领先。
- 高效落地创新应用:借鉴列表中如 TagGPT、KAR 等具体项目的实现思路,成功将 LLM 应用于商品标签自动生成和开放世界推荐,显著提升转化率。
Awesome-LLM-for-RecSys 将原本碎片化的学术研究转化为结构化的工程指南,帮助团队从盲目探索转向高效落地,极大加速了下一代智能推荐系统的研发进程。
运行环境要求
未说明
未说明

快速开始
用于推荐系统的优秀大语言模型资源库 
这是一份关于大语言模型(LLM)相关推荐系统主题的优秀论文和资源合集。
:tada: 我们的综述论文已被**ACM信息管理系统事务(TOIS)**接收:推荐系统如何从大语言模型中获益:综述
:bell: 自从我们的综述论文被归档后,我们将在“1.7 最新研究工作列表”中更新最新的研究成果。
:grin: 我还在微信上每周撰写关于最新LLM增强型推荐系统的论文笔记。欢迎通过扫描二维码关注。
:rocket: 2024年7月9日 - 论文v6发布:我们为TOIS准备的最终定稿版本。
综述论文更新日志
- 2024年7月9日 - 论文v6发布:我们为TOIS准备的最终定稿版本,该版本将被归档。
- 2024年2月5日 - 论文v5发布:全新发布,包含27页主体内容及更详尽的分类体系。
- 2023年6月29日 - 论文v4发布:新增7篇论文。
- 2023年6月28日 - 论文v3发布:修正了错别字。
- 2023年6月12日 - 论文v2发布:在附录中添加了摘要表格。
- 2023年6月9日 - 论文v1发布:初始版本。
1. 论文
我们根据LLM将在推荐系统流程中的哪个环节被应用来进行论文分类,如下图所示。
1.1 LLM用于特征工程
1.1.1 用户与物品级特征增强
| 名称 | 论文 | LLM主干网络(最大规模) | LLM微调策略 | 发表时间 | 链接 |
|---|---|---|---|---|---|
| LLM4KGC | 知识图谱补全模型是少样本学习者:利用大语言模型进行电商关系标注的实证研究 | PaLM (540B)/ ChatGPT | 冻结参数 | Arxiv 2023 | [链接] |
| TagGPT | TagGPT:大语言模型是零样本多模态标签生成器 | ChatGPT | 冻结参数 | Arxiv 2023 | [链接] |
| ICPC | 大语言模型用于用户兴趣旅程建模 | LaMDA (137B) | 全量微调/提示词微调 | Arxiv 2023 | [链接] |
| KAR | 基于大语言模型知识增强的开放世界推荐 | ChatGPT | 冻结参数 | Arxiv 2023 | [链接] |
| PIE | 利用ChatGPT进行产品信息提取 | ChatGPT | 冻结参数 | Arxiv 2023 | [链接] |
| LGIR | 基于LLM的生成对抗网络提升职位推荐效果 | GhatGLM (6B) | 冻结参数 | AAAI 2024 | [链接] |
| GIRL | 基于大语言模型的生成式职位推荐 | BELLE (7B) | 全量微调 | Arxiv 2023 | [链接] |
| LLM-Rec | LLM-Rec:通过提示词引导大语言模型实现个性化推荐 | text-davinci-003 | 冻结参数 | Arxiv 2023 | [链接] |
| HKFR | 异构知识融合:一种基于LLM的个性化推荐新方法 | ChatGPT | 冻结参数 | RecSys 2023 | [链接] |
| LLaMA-E | LLaMA-E:通过多方面指令遵循赋能电商内容创作 | LLaMA (30B) | LoRA | Arxiv 2023 | [链接] |
| EcomGPT | EcomGPT:针对电商场景对大语言模型进行任务链式指令微调 | BLOOMZ (7.1B) | 全量微调 | Arxiv 2023 | [链接] |
| TF-DCon | 利用大语言模型(LLM)赋能无训练数据集浓缩技术,用于基于内容的推荐 | ChatGPT | 冻结参数 | Arxiv 2023 | [链接] |
| RLMRec | 基于大语言模型的推荐表示学习 | ChatGPT | 冻结参数 | WWW 2024 | [链接] |
| LLMRec | LLMRec:结合图增强的大语言模型用于推荐 | ChatGPT | 冻结参数 | WSDM 2024 | [链接] |
| LLMRG | 利用大语言模型推理图提升推荐系统性能 | GPT4 | 冻结参数 | Arxiv 2023 | [链接] |
| CUP | 基于评论文本生成简洁用户画像以进行推荐 | ChatGPT | 冻结参数 | Arxiv 2023 | [链接] |
| SINGLE | 利用大语言模型建模用户浏览路径以进行文章推荐 | ChatGPT | 冻结参数 | Arxiv 2023 | [链接] |
| SAGCN | 先理解再推荐:利用大语言模型进行语义层面的评论挖掘 | Vicuna (13B) | 冻结参数 | Arxiv 2023 | [链接] |
| UEM | 面向个性化语言提示的用户嵌入模型 | FLAN-T5-base (250M) | 全量微调 | Arxiv 2024 | [链接] |
| LLMHG | 基于LLM指导的人本解释性推荐多视角超图学习 | GPT4 | 冻结参数 | Arxiv 2024 | [链接] |
| Llama4Rec | 通过相互增强与适应性聚合将大语言模型融入推荐系统 | LLaMA2 (7B) | 全量微调 | Arxiv 2024 | [链接] |
| LLM4Vis | LLM4Vis:利用ChatGPT进行可解释的可视化推荐 | ChatGPT | 冻结参数 | EMNLP 2023 | [链接] |
| LoRec | LoRec:鲁棒的大语言模型序列推荐,抵御中毒攻击 | LLaMA2 | 冻结参数 | SIGIR 2024 | [链接] |
1.1.2 实例级样本生成
| 名称 | 论文 | 最大规模的LLM主干模型 | LLM微调策略 | 发表时间 | 链接 |
|---|---|---|---|---|---|
| RecInter | 超越静态测试平台:面向动态推荐系统的交互中心型智能体仿真平台 | GPT-4o | 冻结 | EMNLP 2025 | [链接] |
| GReaT | 语言模型是真实的表格数据生成器 | GPT2-medium (355M) | 全量微调 | ICLR 2023 | [链接] |
| ONCE | ONCE:利用开源与闭源大型语言模型提升基于内容的推荐效果 | ChatGPT | 冻结 | WSDM 2024 | [链接] |
| AnyPredict | AnyPredict:用于表格数据预测的基础模型 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| DPLLM | 基于差分隐私大型语言模型进行合成查询生成的隐私保护推荐系统 | T5-XL (3B) | 全量微调 | Arxiv 2023 | [链接] |
| MINT | 大型语言模型增强的叙事驱动型推荐 | text-davinci-003 | 冻结 | RecSys 2023 | [链接] |
| Agent4Rec | 推荐中的生成式智能体研究 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| RecPrompt | RecPrompt:基于大型语言模型的新闻推荐提示词调优框架 | GPT4 | 冻结 | Arxiv 2023 | [链接] |
| PO4ISR | 面向意图驱动会话推荐的大语言模型 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| BEQUE | 淘宝搜索中基于大型语言模型的长尾查询改写 | ChatGLM (6B) | FFT | Arxiv 2023 | [链接] |
| Agent4Ranking | Agent4Ranking:通过多智能体LLM进行个性化查询改写实现语义鲁棒排序 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| PopNudge | 通过偏差分析和语言模型增强的数据增广改进对话式推荐系统 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
1.2 LLM作为特征编码器
1.2.1 表征增强
| 名称 | 论文 | 最大规模的LLM主干模型 | LLM微调策略 | 发表时间 | 链接 |
|---|---|---|---|---|---|
| U-BERT | U-BERT:用于改进推荐的用户表征预训练 | BERT-base (110M) | 全量微调 | AAAI 2021 | [链接] |
| UNBERT | UNBERT:新闻推荐中的用户-新闻匹配BERT | BERT-base (110M) | 全量微调 | IJCAI 2021 | [链接] |
| PLM-NR | 利用预训练语言模型增强新闻推荐 | RoBERTa-base (125M) | 全量微调 | SIGIR 2021 | [链接] |
| Pyramid-ERNIE | 基于预训练语言模型的百度搜索排序 | ERNIE (110M) | 全量微调 | KDD 2021 | [链接] |
| ERNIE-RS | 百度搜索中面向网络规模检索的预训练语言模型 | ERNIE (110M) | 全量微调 | KDD 2021 | [链接] |
| CTR-BERT | CTR-BERT:针对十亿参数教师模型的成本效益知识蒸馏 | 定制BERT (1.5B) | 全量微调 | ENLSP 2021 | [链接] |
| SuKD | 在赞助搜索中为CTR预测学习补充性NLP特征 | RoBERTa-large (355M) | 全量微调 | KDD 2022 | [链接] |
| PREC | 通过即插即用的预训练器提升深度CTR预测以用于新闻推荐 | BERT-base (110M) | 全量微调 | COLING 2022 | [链接] |
| MM-Rec | MM-Rec:视觉语言模型赋能的多模态新闻推荐 | BERT-base (110M) | 全量微调 | SIGIR 2022 | [链接] |
| Tiny-NewsRec | Tiny-NewsRec:高效且有效的基于PLM的新闻推荐 | UniLMv2-base (110M) | 全量微调 | EMNLP 2022 | [链接] |
| PLM4Tag | PTM4Tag:利用预训练模型强化Stack Overflow帖子的标签推荐 | CodeBERT (125M) | 全量微调 | ICPC 2022 | [链接] |
| TwHIN-BERT | TwHIN-BERT:一种社交增强型多语言推文表征预训练语言模型 | BERT-base (110M) | 全量微调 | Arxiv 2022 | [链接] |
| LSH | 使用BERT和查询感知LSH改进非正式文档上的代码示例推荐:一项比较研究 | BERT-base (110M) | 全量微调 | Arxiv 2023 | [链接] |
| LLM2BERT4Rec | 利用大型语言模型进行序列推荐 | text-embedding-ada-002 | 冻结 | RecSys 2023 | [链接] |
| LLM4ARec | 针对推荐中的个性化方面提取对大型语言模型进行提示调优 | GPT2 (110M) | 提示调优 | Arxiv 2023 | [链接] |
| TIGER | 具有生成式检索的推荐系统 | Sentence-T5-base (223M) | 冻结 | NIPS 2023 | [链接] |
| TBIN | TBIN:为CTR预测建模长文本行为数据 | BERT-base (110M) | 冻结 | DLP-RecSys 2023 | [链接] |
| LKPNR | LKPNR:用于个性化新闻推荐框架的LLM和KG | LLaMA2 (7B) | 冻结 | Arxiv 2023 | [链接] |
| SSNA | 朝着高效有效地适应大型语言模型以用于序列推荐的方向发展 | DistilRoBERTa-base (83M) | 分层适配器调优 | Arxiv 2023 | [链接] |
| CollabContext | 协作情境化:弥合协同过滤与预训练语言模型之间的鸿沟 | Instructor-XL (1.5B) | 冻结 | Arxiv 2023 | [链接] |
| LMIndexer | 语言模型作为语义索引器 | T5-base (223M) | 全量微调 | Arxiv 2023 | [链接] |
| Stack | 基于BERT的集成方法用于客户评论的情感分类及其在电子商务助推营销中的应用 | BERT-base (110M) | 冻结 | Arxiv 2023 | [链接] |
| 无 | 利用语言模型进行旅游行程推荐 | BERT-base (110M) | 全量微调 | PMAI@IJCAI 2023 | [链接] |
| UEM | 用于个性化语言提示的用户嵌入模型 | Sentence-T5-base (223M) | 冻结 | Arxiv 2024 | [链接] |
| Social-LLM | Social-LLM:利用语言模型和社交网络数据大规模建模用户行为 | SBERT-MPNet-base (110M) | 冻结 | Arxiv 2024 | [链接] |
| LLMRS | LLMRS:解锁基于LLM的软件购买推荐系统的潜力 | MPNet (110M) | 冻结 | Arxiv 2024 | [链接] |
| KERL | 知识图谱和预训练语言模型增强的对话式推荐系统表征学习 | BERT-mini | 冻结 | TNNLS | [链接] |
| 无 | 利用大型语言模型赋能少样本推荐系统——增强表征 | ChatGPT | 冻结 | IEEE Access | [链接] |
| 无 | 使用语义ID实现更好的泛化:以推荐排序为例 | 未知 | 冻结 | Arxiv 2023 | [链接] |
1.2.2 统一跨领域推荐
| 名称 | 论文 | 最大规模的LLM主干模型 | LLM微调策略 | 发表时间 | 链接 |
|---|---|---|---|---|---|
| ZESRec | 零样本推荐系统 | BERT-base (1.1亿) | 冻结 | Arxiv 2021 | [链接] |
| UniSRec | 向通用序列表示学习迈进:用于推荐系统的BERT-base模型 | BERT-base (1.1亿) | 冻结 | KDD 2022 | [链接] |
| TransRec | TransRec:从多模态反馈中学习可迁移的推荐模型 | BERT-base (1.1亿) | 全量微调 | Arxiv 2022 | [链接] |
| VQ-Rec | 学习向量量化物品表示,用于可迁移的序列推荐系统 | BERT-base (1.1亿) | 冻结 | WWW 2023 | [链接] |
| IDRec vs MoRec | 推荐系统下一步该走向何方?再探基于ID与基于模态的推荐模型 | BERT-base (1.1亿) | 全量微调 | SIGIR 2023 | [链接] |
| TransRec | 探索基于适配器的迁移学习在推荐系统中的应用:实证研究与实践启示 | RoBERTa-base (1.25亿) | 分层适配器微调 | Arxiv 2023 | [链接] |
| TCF | 利用大型语言模型探索基于文本的协同过滤上限:发现与见解 | OPT-175B (1750亿) | 冻结/全量微调 | Arxiv 2023 | [链接] |
| S&R Foundation | 面向冷启动场景的统一搜索与推荐基础模型 | ChatGLM (60亿) | 冻结 | CIKM 2023 | [链接] |
| MISSRec | MISSRec:为推荐系统预训练并迁移多模态兴趣感知序列表示 | CLIP-B/32 (4亿) | 全量微调 | MM 2023 | [链接] |
| UFIN | UFIN:用于多领域点击率预测的通用特征交互网络 | FLAN-T5-base (2.5亿) | 冻结 | Arxiv 2023 | [链接] |
| PMMRec | 对于可迁移的推荐系统来说,多模态就是全部所需 | RoBERTa-large (3.55亿) | 仅对顶层两层进行微调 | ICDE 2024 | [链接] |
| Uni-CTR | 基于大型语言模型的多领域点击率预测统一框架 | Sheared-LLaMA (13亿) | LoRA | Arxiv 2023 | [链接] |
| PCDR | 用于跨领域推荐的提示增强联邦内容表示学习 | BERT-base (1.1亿) | 冻结 | WWW 2024 | [链接] |
1.3 LLM作为评分/排序函数
1.3.1 物品评分任务
| 名称 | 论文 | 最大规模的LLM主干模型 | LLM微调策略 | 发表时间 | 链接 |
|---|---|---|---|---|---|
| LMRecSys | 语言模型作为推荐系统:评估与局限性 | GPT2-XL (1.5B) | 全量微调 | ICBINB 2021 | [链接] |
| PTab | PTab:利用预训练语言模型建模表格数据 | BERT-base (110M) | 全量微调 | Arxiv 2022 | [链接] |
| UniTRec | UniTRec:统一的文本到文本Transformer及联合对比学习框架用于基于文本的推荐 | BART (406M) | 全量微调 | ACL 2023 | [链接] |
| Prompt4NR | 面向新闻推荐的提示学习 | BERT-base (110M) | 全量微调 | SIGIR 2023 | [链接] |
| RecFormer | 文本即一切:为序列化推荐学习语言表征 | LongFormer (149M) | 全量微调 | KDD 2023 | [链接] |
| TabLLM | TabLLM:利用大型语言模型进行少样本表格数据分类 | T0 (11B) | 少样本参数高效微调 | AISTATS 2023 | [链接] |
| Zero-shot GPT | 零样本推荐作为语言建模 | GPT2-medium (355M) | 冻结 | Arxiv 2023 | [链接] |
| FLAN-T5 | 大型语言模型能理解用户偏好吗?在用户评分预测任务上评估大型语言模型 | FLAN-5-XXL (11B) | 全量微调 | Arxiv 2023 | [链接] |
| BookGPT | BookGPT:由大型语言模型赋能的通用图书推荐框架 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| TALLRec | TALLRec:一种有效且高效的微调框架,用于使大型语言模型与推荐任务对齐 | LLaMA (7B) | LoRA | RecSys 2023 | [链接] |
| PBNR | PBNR:基于提示的新闻推荐系统 | T5-small (60M) | 全量微调 | Arxiv 2023 | [链接] |
| CR-SoRec | CR-SoRec:BERT驱动的社会化推荐一致性正则化 | BERT-base (110M) | 全量微调 | RecSys 2023 | [链接] |
| PromptRec | 基于提示的个性化冷启动推荐 | LLaMA (7B) | 冻结 | Arxiv 2023 | [链接] |
| GLRec | 探索大型语言模型在在线职位推荐中理解图数据的能力 | BELLE-LLaMA (7B) | LoRA | Arxiv 2023 | [链接] |
| BERT4CTR | BERT4CTR:一种将预训练语言模型与非文本特征结合以进行CTR预测的有效框架 | RoBERTa-large (355M) | 全量微调 | KDD 2023 | [链接] |
| ReLLa | ReLLa:检索增强型大型语言模型用于推荐中的终身序列行为理解 | Vicuna (13B) | LoRA | WWW 2024 | [链接] |
| TASTE | 文本匹配通过减少流行度偏差提升序列化推荐效果 | T5-base (223M) | 全量微调 | CIKM 2023 | [链接] |
| N/A | 揭示基于文本的推荐系统中的挑战性案例 | BERT-base (110M) | 全量微调 | RecSys研讨会2023 | [链接] |
| ClickPrompt | ClickPrompt:CTR模型是强大的提示生成器,可用于将语言模型适配到CTR预测任务 | RoBERTa-large (355M) | 全量微调 | WWW 2024 | [链接] |
| SetwiseRank | 一种基于集合的零样本排序方法,使用大型语言模型实现高效且高效率的排序 | FLAN-T5-XXL (11B) | 冻结 | Arxiv 2023 | [链接] |
| UPSR | 彻底建模多领域预训练推荐作为语言 | T5-base (223M) | 全量微调 | Arxiv 2023 | [链接] |
| LLM-Rec | 一个模型通吃:大型语言模型是领域无关的推荐系统 | OPT (6.7B) | LoRA | Arxiv 2023 | [链接] |
| LLMRanker | 超越“是”与“否”:通过打细粒度的相关性标签来改进零样本大型语言模型排名器 | FLAN PaLM2 S | 冻结 | Arxiv 2023 | [链接] |
| CoLLM | CoLLM:将协同嵌入整合到大型语言模型中用于推荐 | Vicuna (7B) | LoRA | Arxiv 2023 | [链接] |
| FLIP | FLIP:迈向基于ID的模型与预训练语言模型之间针对CTR预测的细粒度对齐 | RoBERTa-large (355M) | 全量微调 | Arxiv 2023 | [链接] |
| BTRec | BTRec:基于BERT的轨迹推荐用于个性化旅游 | BERT-base (110M) | 全量微调 | Arxiv 2023 | [链接] |
| CLLM4Rec | 协同大型语言模型用于推荐系统 | GPT2 (110M) | 全量微调 | Arxiv 2023 | [链接] |
| CUP | 基于评论文本的简洁用户画像进行推荐 | BERT-base (110M) | 最后一层微调 | Arxiv 2023 | [链接] |
| N/A | 指令蒸馏使大型语言模型成为高效的零样本排名器 | FLAN-T5-XL (3B) | 全量微调 | Arxiv 2023 | [链接] |
| CoWPiRec | 基于词的预训练物品表征用于可迁移的推荐 | BERT-base (110M) | 全量微调 | ICDM 2023 | [链接] |
| RecExplainer | RecExplainer:对齐大型语言模型以提高推荐模型的可解释性 | Vicuna-v1.3 (7B) | LoRA | Arxiv 2023 | [链接] |
| E4SRec | E4SRec:大型语言模型用于序列化推荐的优雅、有效、高效且可扩展的解决方案 | LLaMA2 (13B) | LoRA | Arxiv 2023 | [链接] |
| CER | 推荐自然语言解释中的一致性问题 | GPT2 (110M) | 全量微调 | ECAI 2023 | [链接] |
| LSAT | 大型语言模型推荐系统的增量学习初步研究 | LLaMA (7B) | LoRA | Arxiv 2023 | [链接] |
| Llama4Rec | 通过相互增强和适应性聚合将大型语言模型融入推荐 | LLaMA2 (7B) | 全量微调 | Arxiv 2024 | [链接] |
1.3.2 物品生成任务
| 名称 | 论文 | 最大规模的LLM主干模型 | LLM微调策略 | 发表时间 | 链接 |
|---|---|---|---|---|---|
| GPT4Rec | GPT4Rec: 个性化推荐与用户兴趣解释的生成框架 | GPT2 (1.1亿) | 全量微调 | Arxiv 2023 | [链接] |
| VIP5 | VIP5:面向推荐的多模态基础模型 | T5-base (2.23亿) | 分层适配器微调 | EMNLP 2023 | [链接] |
| P5-ID | 如何为推荐基础模型索引物品ID | T5-small (6千万) | 全量微调 | Arxiv 2023 | [链接] |
| FaiRLLM | ChatGPT对推荐公平吗?大型语言模型推荐中的公平性评估 | ChatGPT | 冻结 | RecSys 2023 | [链接] |
| PALR | PALR:面向推荐的具个性化感知的LLM | LLaMA (70亿) | 全量微调 | Arxiv 2023 | [链接] |
| ChatGPT | 大型语言模型是推荐系统的零样本排序器 | ChatGPT | 冻结 | ECIR 2024 | [链接] |
| AGR | 人工通用推荐系统(AGR)的火花:ChatGPT的早期实验 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| NIR | 基于大型预训练语言模型的零样本下一物品推荐 | GPT3 (1750亿) | 冻结 | Arxiv 2023 | [链接] |
| GPTRec | 使用GPTRec进行生成式序列推荐 | GPT2-medium (3.55亿) | 全量微调 | Gen-IR@SIGIR 2023 | [链接] |
| ChatNews | ChatGPT在新闻推荐中的初步研究:个性化、提供商公平性、假新闻 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| 无 | 大型语言模型在基于语言和物品偏好的近冷启动推荐中具有竞争力 | PaLM (620亿) | 冻结 | RecSys 2023 | [链接] |
| LLMSeqPrompt | 利用大型语言模型进行序列推荐 | OpenAI ada模型 | 微调 | RecSys 2023 | [链接] |
| GenRec | GenRec:用于生成式推荐的大语言模型 | LLaMA (70亿) | LoRA | Arxiv 2023 | [链接] |
| UP5 | UP5:面向公平性的无偏基础模型 | T5-base (2.23亿) | 前缀微调 | Arxiv 2023 | [链接] |
| HKFR | 异质知识融合:通过LLM实现个性化推荐的新方法 | ChatGLM (60亿) | LoRA | RecSys 2023 | [链接] |
| 无 | 大型语言模型的不平等机会:通过职位推荐揭示人口统计学偏见 | ChatGPT | 冻结 | EAAMO 2023 | [链接] |
| BIGRec | 推荐系统中大型语言模型的两步接地范式 | LLaMA (70亿) | LoRA | Arxiv 2023 | [链接] |
| KP4SR | 序列推荐中的知识提示微调 | T5-small (6千万) | 全量微调 | Arxiv 2023 | [链接] |
| RecSysLLM | 利用大型语言模型构建预训练推荐系统 | GLM (100亿) | LoRA | Arxiv 2023 | [链接] |
| POD | 针对高效LLM推荐的提示蒸馏 | T5-small (6千万) | 全量微调 | CIKM 2023 | [链接] |
| 无 | 以严谨方法评估ChatGPT作为推荐系统 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| RaRS | 检索增强型推荐系统:利用大型语言模型提升推荐系统性能 | ChatGPT | 冻结 | RecSys博士生研讨会2023 | [链接] |
| JobRecoGPT | JobRecoGPT——使用LLM的可解释性职位推荐 | GPT4 | 冻结 | Arxiv 2023 | [链接] |
| LANCER | 重新定义序列推荐:通过内容丰富的语言建模学习动态用户兴趣 | GPT2 (1.1亿) | 前缀微调 | Arxiv 2023 | [链接] |
| TransRec | 桥接大型语言模型与推荐的多方面范式 | LLaMA (70亿) | LoRA | Arxiv 2023 | [链接] |
| AgentCF | AgentCF:利用自主语言代理进行推荐系统的协同学习 | text-davinci-003 & gpt-3.5-turbo | 冻结 | WWW 2024 | [链接] |
| P4LM | 利用语言模型和强化学习实现事实性和个性化推荐 | PaLM2-XS | 全量微调 | Arxiv 2023 | [链接] |
| InstructMK | 在推荐系统中融入大型语言模型的多键值策略 | LLaMA (70亿) | 全量微调 | CIKM GenRec 2023 | [链接] |
| LightLM | LightLM:用于生成式推荐的轻量级深而窄的语言模型 | T5-small (6千万) | 全量微调 | Arxiv 2023 | [链接] |
| LlamaRec | LlamaRec:使用大型语言模型进行排序的两阶段推荐 | LLaMA2 (70亿) | QLoRA | PGAI@CIKM 2023 | [链接] |
| 无 | 探索GPT-4V(视觉)的推荐能力:初步案例研究 | GPT-4V | 冻结 | Arxiv 2023 | [链接] |
| 无 | 探索ChatGPT在新闻推荐中的微调 | ChatGPT | gpt-3.5-turbo微调API | Arxiv 2023 | [链接] |
| 无 | LLM在推荐中是否隐性表现出用户歧视?一项实证研究 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| LC-Rec | 通过整合协作语义来调整大型语言模型以用于推荐 | LLaMA (70亿) | LoRA | Arxiv 2023 | [链接] |
| DOKE | 知识插件:增强大型语言模型以实现领域特定的推荐 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| ControlRec | ControlRec:弥合语言模型与个性化推荐之间的语义鸿沟 | T5-base (2.23亿) | 全量微调 | Arxiv 2023 | [链接] |
| LLaRA | LLaRA:大型语言-推荐助手 | LLaMA2 (70亿) | LoRA | SIGIR 2024 | [链接] |
| PO4ISR | 大型语言模型用于意图驱动的会话推荐 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| DRDT | DRDT:基于发散思维的动态反思,用于LLM驱动的序列推荐 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| RecPrompt | RecPrompt:利用大型语言模型进行新闻推荐的提示微调框架 | GPT4 | 冻结 | Arxiv 2023 | [链接] |
| LiT5 | 缩小规模,提升效率:使用Seq2seq编码器-解码器模型进行高效的零样本列表重排序 | T5-XL (30亿) | 全量微调 | Arxiv 2023 | [链接] |
| STELLA | 大型语言模型并非稳定的推荐系统 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| Llama4Rec | 通过相互增强和适应性聚合将大型语言模型融入推荐 | LLaMA2 (70亿) | 全量微调 | Arxiv 2024 | [链接] |
| RECLLM | 理解基于ChatGPT的推荐系统中的偏见:提供商公平性、时间稳定性及近期性 | ChatGPT | 冻结 | Arxiv 2024 | [链接] |
| DEALRec | 面向LLM推荐的数据高效微调 | LLaMA (70亿) | LoRA | Arxiv 2024 | [链接] |
1.3.3 混合任务
| 名称 | 论文 | 最大基础大模型 | 大模型微调策略 | 发表时间 | 链接 |
|---|---|---|---|---|---|
| P5 | 推荐即语言处理(RLP):统一的预训练、个性化提示与预测范式(P5) | T5-base (2.23亿) | 全量微调 | RecSys 2022 | [链接] |
| M6-Rec | M6-Rec:生成式预训练语言模型是开放式推荐系统 | M6-base (3亿) | 选项微调 | Arxiv 2022 | [链接] |
| InstructRec | 推荐即指令遵循:一种由大型语言模型赋能的推荐方法 | FLAN-T5-XL (30亿) | 全量微调 | Arxiv 2023 | [链接] |
| ChatGPT | ChatGPT 是一个好的推荐系统吗?一项初步研究 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| ChatGPT | ChatGPT 在搜索方面表现如何?探究大型语言模型作为重排序代理 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| ChatGPT | 揭示 ChatGPT 在推荐系统中的能力 | ChatGPT | 冻结 | RecSys 2023 | [链接] |
| BDLM | 桥接领域特定模型与通用大模型之间的信息鸿沟,用于个性化推荐 | Vicuna (70亿) | 全量微调 | Arxiv 2023 | [链接] |
| RecRanker | RecRanker:将大型语言模型通过指令微调用作 Top-k 推荐的排序器 | LLaMA2 (130亿) | 全量微调 | Arxiv 2023 | [链接] |
1.4 大模型用于用户交互
1.4.1 任务导向型用户交互
| 名称 | 论文 | 最大基础大模型 | 大模型微调策略 | 发表时间 | 链接 |
|---|---|---|---|---|---|
| TG-ReDial | 走向主题引导的对话式推荐系统 | BERT-base (1.1亿) & GPT2 (1.1亿) | 未知 | COLING 2020 | [链接] |
| TCP | 跟我来:面向目标驱动的推荐对话系统的对话规划 | BERT-base (1.1亿) | 全量微调 | Arxiv 2022 | [链接] |
| MESE | 利用上下文感知的商品元信息提升对话式推荐系统的质量 | DistilBERT (6700万) & GPT2 (1.1亿) | 全量微调 | ACL 2022 | [链接] |
| UniMIND | 面向多目标对话式推荐系统的统一多任务学习框架 | BART-base (1.39亿) | 全量微调 | ACM TOIS 2023 | [链接] |
| VRICR | 针对对话式推荐的不完全知识图谱上的变分推理 | BERT-base (1.1亿) | 全量微调 | WSDM 2023 | [链接] |
| KECR | 对话式推荐中的显式知识图谱推理 | BERT-base (1.1亿) & GPT2 (1.1亿) | 冻结 | ACM TIST 2023 | [链接] |
| 无 | 大型语言模型作为零样本对话式推荐系统 | GPT4 | 冻结 | CIKM 2023 | [链接] |
| MuseChat | MuseChat:一个面向视频的对话式音乐推荐系统 | Vicuna (70亿) | LoRA | Arxiv 2023 | [链接] |
| 无 | 电商售前对话中,对话式推荐系统与大型语言模型天生一对 | Chinese-Alpaca (70亿) | LoRA | EMNLP 2023 Findings | [链接] |
| 无 | ChatGPT 用于对话式推荐:通过反馈重新提示来优化推荐 | ChatGPT | 冻结 | Arxiv 2024 | [链接] |
1.4.2 开放式用户交互
| 名称 | 论文 | 最大基础大模型 | 大模型微调策略 | 发表时间 | 链接 |
|---|---|---|---|---|---|
| BARCOR | BARCOR:迈向对话式推荐系统的统一框架 | BART-base (1.39亿) | 选择性层微调 | Arxiv 2022 | [链接] |
| RecInDial | RecInDial:基于预训练语言模型的对话式推荐统一框架 | DialoGPT (1.1亿) | 全量微调 | AACL 2022 | [链接] |
| UniCRS | 通过知识增强的提示学习迈向统一的对话式推荐系统 | DialoGPT-small (1.76亿) | 冻结 | KDD 2022 | [链接] |
| T5-CR | 多任务端到端训练提升对话式推荐效果 | T5-base (2.23亿) | 全量微调 | Arxiv 2023 | [链接] |
| TtW | 言行一致:用于对话式音乐推荐的合成数据生成 | T5-base (2.23亿) & T5-XXL (110亿) | 全量微调和冻结 | Arxiv 2023 | [链接] |
| 无 | 重新思考大型语言模型时代下对话式推荐的评估 | ChatGPT | 冻结 | EMNLP 2023 | [链接] |
| PECRS | 参数高效的对话式推荐系统作为语言处理任务 | GPT2-medium (3.55亿) | LoRA | EACL 2024 | [链接] |
1.5 大模型用于推荐系统流水线控制器
| 名称 | 论文 | 最大基础大模型 | 大模型微调策略 | 发表时间 | 链接 |
|---|---|---|---|---|---|
| Chat-REC | Chat-REC:迈向交互式且可解释的大模型增强型推荐系统 | ChatGPT | 冻结 | Arxiv 2023 | [链接] |
| RecLLM | 在对话式推荐系统中利用大型语言模型 | LLaMA (70亿) | 全量微调 | Arxiv 2023 | [链接] |
| RAH | RAH!推荐系统助手—人类:以人类为中心的大型语言模型推荐框架 | GPT4 | 冻结 | Arxiv 2023 | [链接] |
| RecMind | RecMind:大型语言模型驱动的推荐智能体 | ChatGPT | 冻结 | NAACL 2024 | [链接] |
| InteRecAgent | 推荐AI智能体:整合大型语言模型实现交互式推荐 | GPT4 | 冻结 | Arxiv 2023 | [链接] |
| CORE | 为推荐系统插上对话智能体的翅膀 | 无 | 无 | NIPS 2023 | [链接] |
| LLMCRS | 一种大型语言模型增强的对话式推荐系统 | LLaMA (70亿) | 全量微调 | Arxiv 2023 | [链接] |
1.6 相关综述论文
| 论文 | 发表刊物 | 链接 |
|---|---|---|
| GR-LLMs:基于大语言模型的生成式推荐最新进展 | Arxiv 2025 | [链接] |
| 未来是智能体驱动的:多智能体推荐系统的定义、视角与开放挑战 | Arxiv 2025 | [链接] |
| 基础模型赋能的推荐系统综述:从特征型、生成式到智能体范式 | Arxiv 2025 | [链接] |
| 个性化综述:从RAG到智能体 | Arxiv 2025 | [链接] |
| 大语言模型赋能的推荐与搜索智能体综述:迈向下一代信息检索 | Arxiv 2025 | [链接] |
| 以智能体为中心的信息获取 | Arxiv 2025 | [链接] |
| 基于LLM的新闻推荐系统综述 | Arxiv 2025 | [链接] |
| 推荐系统中LLM赋能的智能体综述 | Arxiv 2025 | [链接] |
| 面向大语言模型时代的冷启动推荐:全面综述与路线图 | Arxiv 2025 | [链接] |
| 大语言模型增强的推荐系统:分类、趋势、应用与未来 | Arxiv 2024 | [链接] |
| 大语言模型智能体时代的推荐系统:综述 | 预印本 | [链接] |
| 推荐领域大语言模型高效解决方案综述 | Arxiv 2024 | [链接] |
| 迈向新一代LLM驱动的推荐系统:综述及展望 | Arxiv 2024 | [链接] |
| 信息检索系统中的偏见与不公平:LLM时代的新挑战 | KDD 2024 | [链接] |
| 条条大路通罗马:揭示推荐系统在LLM时代的发展轨迹 | Arxiv 2024 | [链接] |
| 社交和电商推荐系统中生成式AI落地的行业视角综述 | Arxiv 2024 | [链接] |
| 大语言模型时代生成式搜索与推荐综述 | Arxiv 2024 | [链接] |
| 搜索引擎服务与大语言模型的结合:愿景与挑战 | Arxiv 2024 | [链接] |
| 使用生成模型的现代推荐系统综述(Gen-RecSys) | Arxiv 2024 | [链接] |
| 探讨大语言模型对推荐系统的影响:广泛综述 | Arxiv 2024 | [链接] |
| 推荐系统的基础模型:综述与新视角 | Arxiv 2024 | [链接] |
| 针对推荐系统的提示工程:综合框架与实证分析 | Arxiv 2024 | [链接] |
| 大语言模型时代的用户建模:当前研究与未来方向 | IEEE数据工程简报 2023 | [链接] |
| 大语言模型用于个性化和可解释性推荐的综述 | Arxiv 2023 | [链接] |
| 大语言模型用于生成式推荐:综述与前瞻性讨论 | Arxiv 2023 | [链接] |
| 大语言模型用于信息检索:综述 | Arxiv 2023 | [链接] |
| 当大语言模型与个性化相遇时:挑战与机遇的视角 | Arxiv 2023 | [链接] |
| 大语言模型时代的推荐系统 | Arxiv 2023 | [链接] |
| 大语言模型用于推荐的综述 | Arxiv 2023 | [链接] |
| 预训练、提示与推荐:推荐系统中语言建模范式的适应性综合综述 | TACL 2023 | [链接] |
| 推荐系统中的自监督学习:综述 | TKDE 2022 | [链接] |
1.7 最新研究工作列表
| Paper | Publication | Link |
|---|---|---|
| Large Language Model Can Interpret Latent Space of Sequential Recommender | Arxiv 2023 | [Link] |
| Zero-Shot Recommendations with Pre-Trained Large Language Models for Multimodal Nudging | Arxiv 2023 | [Link] |
| INTERS: Unlocking the Power of Large Language Models in Search with Instruction Tuning | Arxiv 2024 | [Link] |
| Evaluation of Synthetic Datasets for Conversational Recommender Systems | Arxiv 2023 | [Link] |
| Generative Recommendation: Towards Next-generation Recommender Paradigm | Arxiv 2023 | [Link] |
| Towards Personalized Prompt-Model Retrieval for Generative Recommendation | Arxiv 2023 | [Link] |
| Generative Next-Basket Recommendation | RecSys 2023 | [Link] |
| Unlocking the Potential of Large Language Models for Explainable Recommendations | Arxiv 2023 | [Link] |
| Logic-Scaffolding: Personalized Aspect-Instructed Recommendation Explanation Generation using LLMs | Falcon (40B) | Frozen |
| Improving Sequential Recommendations with LLMs | Arxiv 2024 | [Link] |
| A Multi-Agent Conversational Recommender System | Arxiv 2024 | [Link] |
| TransFR: Transferable Federated Recommendation with Pre-trained Language Models | Arxiv 2024 | [Link] |
| Large Language Model Distilling Medication Recommendation Model | Arxiv 2024 | [Link] |
| Uncertainty-Aware Explainable Recommendation with Large Language Models | Arxiv 2024 | [Link] |
| Natural Language User Profiles for Transparent and Scrutable Recommendations | Arxiv 2024 | [Link] |
| Leveraging LLMs for Unsupervised Dense Retriever Ranking | Arxiv 2024 | [Link] |
| RA-Rec: An Efficient ID Representation Alignment Framework for LLM-based Recommendation | Arxiv 2024 | [Link] |
| A Multi-Agent Conversational Recommender System | Arxiv 2024 | [Link] |
| Fairly Evaluating Large Language Model-based Recommendation Needs Revisit the Cross-Entropy Loss | Arxiv 2024 | [Link] |
| SearchAgent: A Lightweight Collaborative Search Agent with Large Language Models | Arxiv 2024 | [Link] |
| Large Language Model Interaction Simulator for Cold-Start Item Recommendation | Arxiv 2024 | [Link] |
| Enhancing ID and Text Fusion via Alternative Training in Session-based Recommendation | Arxiv 2024 | [Link] |
| eCeLLM: Generalizing Large Language Models for E-commerce from Large-scale, High-quality Instruction Data | Arxiv 2024 | [Link] |
| LLM-Enhanced User-Item Interactions: Leveraging Edge Information for Optimized Recommendations | Arxiv 2024 | [Link] |
| LLM-based Federated Recommendation | Arxiv 2024 | [Link] |
| Rethinking Large Language Model Architectures for Sequential Recommendations | Arxiv 2024 | [Link] |
| Large Language Model with Graph Convolution for Recommendation | Arxiv 2024 | [Link] |
| Rec-GPT4V: Multimodal Recommendation with Large Vision-Language Models | Arxiv 2024 | [Link] |
| Enhancing Recommendation Diversity by Re-ranking with Large Language Models | Arxiv 2024 | [Link] |
| Are ID Embeddings Necessary? Whitening Pre-trained Text Embeddings for Effective Sequential Recommendation | Arxiv 2024 | [Link] |
| SPAR: Personalized Content-Based Recommendation via Long Engagement Attention | Arxiv 2024 | [Link] |
| Cognitive Personalized Search Integrating Large Language Models with an Efficient Memory Mechanism | WWW 2024 | [Link] |
| Large Language Models as Data Augmenters for Cold-Start Item Recommendation | Arxiv 2024 | [Link] |
| Explain then Rank: Scale Calibration of Neural Rankers Using Natural Language Explanations from Large Language Models | Arxiv 2024 | [Link] |
| LLM4SBR: A Lightweight and Effective Framework for Integrating Large Language Models in Session-based Recommendation | Arxiv 2024 | [Link] |
| Breaking the Barrier: Utilizing Large Language Models for Industrial Recommendation Systems through an Inferential Knowledge Graph | Arxiv 2024 | [Link] |
| User-LLM: Efficient LLM Contextualization with User Embeddings | Arxiv 2024 | [Link] |
| Stealthy Attack on Large Language Model based Recommendation | Arxiv 2024 | [Link] |
| Multi-Agent Collaboration Framework for Recommender Systems | Arxiv 2024 | [Link] |
| Item-side Fairness of Large Language Model-based Recommendation System | WWW 2024 | [Link] |
| Integrating Large Language Models with Graphical Session-Based Recommendation | Arxiv 2024 | [Link] |
| Language-Based User Profiles for Recommendation | LLM-IGS@WSDM2024 | [Link] |
| BASES: Large-scale Web Search User Simulation with Large Language Model based Agents | Arxiv 2024 | [Link] |
| Prospect Personalized Recommendation on Large Language Model-based Agent Platform | Arxiv 2024 | [Link] |
| Sequence-level Semantic Representation Fusion for Recommender Systems | Arxiv 2024 | [Link] |
| Corpus-Steered Query Expansion with Large Language Models | ECAL 2024 | [Link] |
| NoteLLM: A Retrievable Large Language Model for Note Recommendation | WWW 2024 | [Link] |
| An Interpretable Ensemble of Graph and Language Models for Improving Search Relevance in E-Commerce | WWW 2024 | [Link] |
| LLM-Ensemble: Optimal Large Language Model Ensemble Method for E-commerce Product Attribute Value Extraction | Arxiv 2024 | [Link] |
| Enhancing Long-Term Recommendation with Bi-level Learnable Large Language Model Planning | Arxiv 2024 | [Link] |
| InteraRec: Interactive Recommendations Using Multimodal Large Language Models | Arxiv 2024 | [Link] |
| ChatDiet: Empowering Personalized Nutrition-Oriented Food Recommender Chatbots through an LLM-Augmented Framework | CHASE 2024 | [Link] |
| Towards Efficient and Effective Unlearning of Large Language Models for Recommendation | Arxiv 2024 | [Link] |
| Generative News Recommendation | WWW 2024 | [Link] |
| Bridging Language and Items for Retrieval and Recommendation | Arxiv 2024 | [Link] |
| Can Small Language Models be Good Reasoners for Sequential Recommendation? | WWW 2024 | [Link] |
| Aligning Large Language Models for Controllable Recommendations | Arxiv 2024 | [Link] |
| Personalized Audiobook Recommendations at Spotify Through Graph Neural Networks | WWW 2024 | [Link] |
| Towards Graph Foundation Models for Personalization | Arxiv 2024 | [Link] |
| CFaiRLLM: Consumer Fairness Evaluation in Large-Language Model Recommender System | Arxiv 2024 | [Link] |
| CoRAL: Collaborative Retrieval-Augmented Large Language Models Improve Long-tail Recommendation | Arxiv 2024 | [Link] |
| RecAI: Leveraging Large Language Models for Next-Generation Recommender Systems | WWW 2024 Demo | [Link] |
| KELLMRec: Knowledge-Enhanced Large Language Models for Recommendation | Arxiv 2024 | [Link] |
| USimAgent: Large Language Models for Simulating Search Users | Arxiv 2024 | [Link] |
| CALRec: Contrastive Alignment of Generative LLMs For Sequential Recommendation | Arxiv 2024 | [Link] |
| Integrating Large Language Models with Graphical Session-Based Recommendation | Arxiv 2024 | [Link] |
| Language-Based User Profiles for Recommendation | LLM-IGS@WSDM2024 | [Link] |
| BASES: Large-scale Web Search User Simulation with Large Language Model based Agents | Arxiv 2024 | [Link] |
| Prospect Personalized Recommendation on Large Language Model-based Agent Platform | Arxiv 2024 | [Link] |
| Sequence-level Semantic Representation Fusion for Recommender Systems | Arxiv 2024 | [Link] |
| Corpus-Steered Query Expansion with Large Language Models | EACL 2024 | [Link] |
| NoteLLM: A Retrievable Large Language Model for Note Recommendation | WWW 2024 | [Link] |
| An Interpretable Ensemble of Graph and Language Models for Improving Search Relevance in E-Commerce | WWW 2024 | [Link] |
| LLM-Ensemble: Optimal Large Language Model Ensemble Method for E-commerce Product Attribute Value Extraction | SIGIR 2024 | [Link] |
| Enhancing Long-Term Recommendation with Bi-level Learnable Large Language Model Planning | SIGIR 2024 | [Link] |
| Towards Efficient and Effective Unlearning of Large Language Models for Recommendation | FCS | [Link] |
| Generative News Recommendation | WWW 2024 | [Link] |
| Bridging Language and Items for Retrieval and Recommendation | Arxiv 2024 | [Link] |
| Can Small Language Models be Good Reasoners for Sequential Recommendation? | WWW 2024 | [Link] |
| Aligning Large Language Models for Controllable Recommendations | Arxiv 2024 | [Link] |
| Personalized Audiobook Recommendations at Spotify Through Graph Neural Networks | WWW 2024 | [Link] |
| CFaiRLLM: Consumer Fairness Evaluation in Large-Language Model Recommender System | Arxiv 2024 | [Link] |
| CoRAL: Collaborative Retrieval-Augmented Large Language Models Improve Long-tail Recommendation | Arxiv 2024 | [Link] |
| RecAI: Leveraging Large Language Models for Next-Generation Recommender Systems | WWW 2024 | [Link] |
| KELLMRec: Knowledge-Enhanced Large Language Models for Recommendation | Arxiv 2024 | [Link] |
| Towards Graph Foundation Models for Personalization | Arxiv 2024 | [Link] |
| USimAgent: Large Language Models for Simulating Search Users | Arxiv 2024 | [Link] |
| The Whole is Better than the Sum: Using Aggregated Demonstrations in In-Context Learning for Sequential Recommendation | NAACL 2024 | [Link] |
| PPM : A Pre-trained Plug-in Model for Click-through Rate Prediction | WWW 2024 | [Link] |
| Evaluating Large Language Models as Generative User Simulators for Conversational Recommendation | Arxiv 2024 | [Link] |
| Towards Unified Multi-Modal Personalization: Large Vision-Language Models for Generative Recommendation and Beyond | ICLR 2024 | [Link] |
| Harnessing Large Language Models for Text-Rich Sequential Recommendation | Arxiv 2024 | [Link] |
| A Large Language Model Enhanced Sequential Recommender for Joint Video and Comment Recommendation | Arxiv 2024 | [Link] |
| Could Small Language Models Serve as Recommenders? Towards Data-centric Cold-start Recommendations | Arxiv 2024 | [Link] |
| Play to Your Strengths: Collaborative Intelligence of Conventional Recommender Models and Large Language Models | Arxiv 2024 | [Link] |
| Reinforcement Learning-based Recommender Systems with Large Language Models for State Reward and Action Modeling | Arxiv 2024 | [Link] |
| Large Language Models Enhanced Collaborative Filtering | Arxiv 2024 | [Link] |
| Improving Content Recommendation: Knowledge Graph-Based Semantic Contrastive Learning for Diversity and Cold-Start Users | LREC-COLING 2024 | [Link] |
| Sequential Recommendation with Latent Relations based on Large Language Model | Arxiv 2024 | [Link] |
| Enhanced Generative Recommendation via Content and Collaboration Integration | Arxiv 2024 | [Link] |
| To Recommend or Not: Recommendability Identification in Conversations with Pre-trained Language Models | Arxiv 2024 | [Link] |
| IDGenRec: LLM-RecSys Alignment with Textual ID Learning | SIGIR 2024 | [Link] |
| Breaking the Length Barrier: LLM-Enhanced CTR Prediction in Long Textual User Behaviors | SIGIR 2024 | [Link] |
| Make Large Language Model a Better Ranker | Arxiv 2024 | [Link] |
| Do Large Language Models Rank Fairly? An Empirical Study on the Fairness of LLMs as Rankers | NAACL 2024 | [Link] |
| IISAN: Efficiently Adapting Multimodal Representation for Sequential Recommendation with Decoupled PEFT | SIGIR 2024 | [Link] |
| Where to Move Next: Zero-shot Generalization of LLMs for Next POI Recommendation | Arxiv 2024 | [Link] |
| Tired of Plugins? Large Language Models Can Be End-To-End Recommender | Arxiv 2024 | [Link] |
| Aligning Large Language Models with Recommendation Knowledge | NAACL 2024 | [Link] |
| Enhancing Content-based Recommendation via Large Language Model | Arxiv 2024 | [Link] |
| DRE: Generating Recommendation Explanations by Aligning Large Language Models at Data-level | Arxiv 2024 | [Link] |
| Optimization Methods for Personalizing Large Language Models through Retrieval Augmentation | Arxiv 2024 | [Link] |
| Q-PEFT: Query-dependent Parameter Efficient Fine-tuning for Text Reranking with Large Language Models | Arxiv 2024 | [Link] |
| JobFormer: Skill-Aware Job Recommendation with Semantic-Enhanced Transformer | Arxiv 2024 | [Link] |
| PMG : Personalized Multimodal Generation with Large Language Models | Arxiv 2024 | [Link] |
| The Elephant in the Room: Rethinking the Usage of Pre-trained Language Model in Sequential Recommendation | Arxiv 2024 | [Link] |
| Exact and Efficient Unlearning for Large Language Model-based Recommendation | Arxiv 2024 | [Link] |
| Large Language Models meet Collaborative Filtering: An Efficient All-round LLM-based Recommender System | Arxiv 2024 | [Link] |
| Behavior Alignment: A New Perspective of Evaluating LLM-based Conversational Recommendation Systems | SIGIR 2024 | [Link] |
| Generating Diverse Criteria On-the-Fly to Improve Point-wise LLM Rankers | Arxiv 2024 | [Link] |
| RecGPT: Generative Personalized Prompts for Sequential Recommendation via ChatGPT Training Paradigm | Arxiv 2024 | [Link] |
| MMGRec: Multimodal Generative Recommendation with Transformer Model | Arxiv 2024 | [Link] |
| Hi-Gen: Generative Retrieval For Large-Scale Personalized E-commerce Search | Arxiv 2024 | [Link] |
| Contrastive Quantization based Semantic Code for Generative Recommendation | Arxiv 2024 | [Link] |
| ImplicitAVE: An Open-Source Dataset and Multimodal LLMs Benchmark for Implicit Attribute Value Extraction | Arxiv 2024 | [Link] |
| Large Language Models for Next Point-of-Interest Recommendation | SIGIR 2024 | [Link] |
| Ranked List Truncation for Large Language Model-based Re-Ranking | SIGIR 2024 | [Link] |
| Large Language Models as Conversational Movie Recommenders: A User Study | Arxiv 2024 | [Link] |
| Distillation Matters: Empowering Sequential Recommenders to Match the Performance of Large Language Model | Arxiv 2024 | [Link] |
| Efficient and Responsible Adaptation of Large Language Models for Robust Top-k Recommendations | Arxiv 2024 | [Link] |
| FairEvalLLM. A Comprehensive Framework for Benchmarking Fairness in Large Language Model Recommender Systems | Arxiv 2024 | [Link] |
| Improve Temporal Awareness of LLMs for Sequential Recommendation | Arxiv 2024 | [Link] |
| CALRec: Contrastive Alignment of Generative LLMs For Sequential Recommendation | Arxiv 2024 | [Link] |
| Knowledge Adaptation from Large Language Model to Recommendation for Practical Industrial Application | Arxiv 2024 | [Link] |
| DynLLM: When Large Language Models Meet Dynamic Graph Recommendation | Arxiv 2024 | [Link] |
| Learnable Tokenizer for LLM-based Generative Recommendation | Arxiv 2024 | [Link] |
| CELA: Cost-Efficient Language Model Alignment for CTR Prediction | Arxiv 2024 | [Link] |
| RDRec: Rationale Distillation for LLM-based Recommendation | ACL 2024 | [Link] |
| EmbSum: Leveraging the Summarization Capabilities of Large Language Models for Content-Based Recommendations | Arxiv 2024 | [Link] |
| Reindex-Then-Adapt: Improving Large Language Models for Conversational Recommendation | Arxiv 2024 | [Link] |
| RecGPT: Generative Pre-training for Text-based Recommendation | ACL 2024 | [Link] |
| Let Me Do It For You: Towards LLM Empowered Recommendation via Tool Learning | SIGIR 2024 | [Link] |
| Finetuning Large Language Model for Personalized Ranking | Arxiv 2024 | [Link] |
| LLMs for User Interest Exploration: A Hybrid Approach | Arxiv 2024 | [Link] |
| NoteLLM-2: Multimodal Large Representation Models for Recommendation | Arxiv 2024 | [Link] |
| Multimodality Invariant Learning for Multimedia-Based New Item Recommendation | Arxiv 2024 | [Link] |
| SLMRec: Empowering Small Language Models for Sequential Recommendation | Arxiv 2024 | [Link] |
| Keyword-driven Retrieval-Augmented Large Language Models for Cold-start User Recommendations | Arxiv 2024 | [Link] |
| Generating Query Recommendations via LLMs | Arxiv 2024 | [Link] |
| Large Language Models Enhanced Sequential Recommendation for Long-tail User and Item | Arxiv 2024 | [Link] |
| DisCo: Towards Harmonious Disentanglement and Collaboration between Tabular and Semantic Space for Recommendation | KDD 2024 | [Link] |
| LLM-RankFusion: Mitigating Intrinsic Inconsistency in LLM-based Ranking | Arxiv 2024 | [Link] |
| A Practice-Friendly Two-Stage LLM-Enhanced Paradigm in Sequential Recommendation | Arxiv 2024 | [Link] |
| Large Language Models as Recommender Systems: A Study of Popularity Bias | Gen-IR@SIGIR24 | [Link] |
| Privacy in LLM-based Recommendation: Recent Advances and Future Directions | Arxiv 2024 | [Link] |
| An LLM-based Recommender System Environment | Arxiv 2024 | [Link] |
| Robust Interaction-based Relevance Modeling for Online E-Commerce and LLM-based Retrieval | ECML-PKDD 2024 | [Link] |
| Large Language Models Make Sample-Efficient Recommender Systems | FCS | [Link] |
| XRec: Large Language Models for Explainable Recommendation | Arxiv 2024 | [Link] |
| Exploring User Retrieval Integration towards Large Language Models for Cross-Domain Sequential Recommendation | Arxiv 2024 | [Link] |
| Large Language Models as Evaluators for Recommendation Explanations | Arxiv 2024 | [Link] |
| Text-like Encoding of Collaborative Information in Large Language Models for Recommendation | ACL 2024 | [Link] |
| Item-Language Model for Conversational Recommendation | Arxiv 2024 | [Link] |
| Improving LLMs for Recommendation with Out-Of-Vocabulary Tokens | Arxiv 2024 | [Link] |
| On Softmax Direct Preference Optimization for Recommendation | Arxiv 2024 | [Link] |
| TokenRec: Learning to Tokenize ID for LLM-based Generative Recommendation | Arxiv 2024 | [Link] |
| DELRec: Distilling Sequential Pattern to Enhance LLM-based Recommendation | Arxiv 2024 | [Link] |
| TourRank: Utilizing Large Language Models for Documents Ranking with a Tournament-Inspired Strategy | Arxiv 2024 | [Link] |
| Multi-Layer Ranking with Large Language Models for News Source Recommendation | SIGIR 2024 | [Link] |
| Intermediate Distillation: Data-Efficient Distillation from Black-Box LLMs for Information Retrieval | Arxiv 2024 | [Link] |
| LLM-enhanced Reranking in Recommender Systems | Arxiv 2024 | [Link] |
| LLM4MSR: An LLM-Enhanced Paradigm for Multi-Scenario Recommendation | Arxiv 2024 | [Link] |
| Taxonomy-Guided Zero-Shot Recommendations with LLMs | Arxiv 2024 | [Link] |
| EAGER: Two-Stream Generative Recommender with Behavior-Semantic Collaboration | KDD 2024 | [Link] |
| An Investigation of Prompt Variations for Zero-shot LLM-based Rankers | Arxiv 2024 | [Link] |
| Optimizing Novelty of Top-k Recommendations using Large Language Models and Reinforcement Learning | KDD 2024 | [Link] |
| Enhancing Collaborative Semantics of Language Model-Driven Recommendations via Graph-Aware Learning | Arxiv 2024 | [Link] |
| Decoding Matters: Addressing Amplification Bias and Homogeneity Issue for LLM-based Recommendation | Arxiv 2024 | [Link] |
| FIRST: Faster Improved Listwise Reranking with Single Token Decoding | Arxiv 2024 | [Link] |
| LLM-Powered Explanations: Unraveling Recommendations Through Subgraph Reasoning | Arxiv 2024 | [Link] |
| DemoRank: Selecting Effective Demonstrations for Large Language Models in Ranking Task | Arxiv 2024 | [Link] |
| ELCoRec: Enhance Language Understanding with Co-Propagation of Numerical and Categorical Features for Recommendation | Arxiv 2024 | [Link] |
| Generative Explore-Exploit: Training-free Optimization of Generative Recommender Systems using LLM Optimizers | ACL 2024 | [Link] |
| ProductAgent: Benchmarking Conversational Product Search Agent with Asking Clarification Questions | Arxiv 2024 | [Link] |
| MemoCRS: Memory-enhanced Sequential Conversational Recommender Systems with Large Language Models | Arxiv 2024 | [Link] |
| Preference Distillation for Personalized Generative Recommendation | Arxiv 2024 | [Link] |
| Towards Bridging the Cross-modal Semantic Gap for Multi-modal Recommendation | Arxiv 2024 | [Link] |
| Language Models Encode Collaborative Signals in Recommendation | Arxiv 2024 | [Link] |
| A Neural Matrix Decomposition Recommender System Model based on the Multimodal Large Language Model | Arxiv 2024 | [Link] |
| LLMGR: Large Language Model-based Generative Retrieval in Alipay Search | SIGIR 2024 | [Link] |
| Enhancing Sequential Recommenders with Augmented Knowledge from Aligned Large Language Models | SIGIR 2024 | [Link] |
| Reinforced Prompt Personalization for Recommendation with Large Language Models | Arxiv 2024 | [Link] |
| Improving Retrieval in Sponsored Search by Leveraging Query Context Signals | Arxiv 2024 | [Link] |
| Generative Retrieval with Preference Optimization for E-commerce Search | Arxiv 2024 | [Link] |
| GenRec: Generative Personalized Sequential Recommendation | Arxiv 2024 | [Link] |
| Breaking the Hourglass Phenomenon of Residual Quantization: Enhancing the Upper Bound of Generative Retrieval | Arxiv 2024 | [Link] |
| Enhancing Taobao Display Advertising with Multimodal Representations: Challenges, Approaches and Insights | CIKM 2024 | [Link] |
| Leveraging LLM Reasoning Enhances Personalized Recommender Systems | ACL 2024 | [Link] |
| Multi-Aspect Reviewed-Item Retrieval via LLM Query Decomposition and Aspect Fusion | Arxiv 2024 | [Link] |
| Lifelong Personalized Low-Rank Adaptation of Large Language Models for Recommendation | Arxiv 2024 | [Link] |
| Exploring Query Understanding for Amazon Product Search | Arxiv 2024 | [Link] |
| A Decoding Acceleration Framework for Industrial Deployable LLM-based Recommender Systems | Arxiv 2024 | [Link] |
| Prompt Tuning as User Inherent Profile Inference Machine | Arxiv 2024 | [Link] |
| Beyond Inter-Item Relations: Dynamic Adaptive Mixture-of-Experts for LLM-Based Sequential Recommendation | Arxiv 2024 | [Link] |
| Review-driven Personalized Preference Reasoning with Large Language Models for Recommendation | Arxiv 2024 | [Link] |
| DaRec: A Disentangled Alignment Framework for Large Language Model and Recommender System | Arxiv 2024 | [Link] |
| LLM4DSR: Leveraing Large Language Model for Denoising Sequential Recommendation | Arxiv 2024 | [Link] |
| EasyRec: Simple yet Effective Language Models for Recommendation | Arxiv 2024 | [Link] |
| Collaborative Cross-modal Fusion with Large Language Model for Recommendation | CIKM 2024 | [Link] |
| Customizing Language Models with Instance-wise LoRA for Sequential Recommendation | Arxiv 2024 | [Link] |
| Efficient and Deployable Knowledge Infusion for Open-World Recommendations via Large Language Models | Arxiv 2024 | [Link] |
| CoRA: Collaborative Information Perception by Large Language Model's Weights for Recommendation | Arxiv 2024 | [Link] |
| GANPrompt: Enhancing Robustness in LLM-Based Recommendations with GAN-Enhanced Diversity Prompts | Arxiv 2024 | [Link] |
| Harnessing Multimodal Large Language Models for Multimodal Sequential Recommendation | Arxiv 2024 | [Link] |
| DLCRec: A Novel Approach for Managing Diversity in LLM-Based Recommender Systems | Arxiv | [Link] |
| LARR: Large Language Model Aided Real-time Scene Recommendation with Semantic Understanding | RecSys 2024 | [Link] |
| SC-Rec: Enhancing Generative Retrieval with Self-Consistent Reranking for Sequential Recommendation | Arxiv 2024 | [Link] |
| Are LLM-based Recommenders Already the Best? Simple Scaled Cross-entropy Unleashes the Potential of Traditional Sequential Recommenders | Arxiv 2024 | [Link] |
| HRGraph: Leveraging LLMs for HR Data Knowledge Graphs with Information Propagation-based Job Recommendation | KaLLM 2024 | [Link] |
| An Extremely Data-efficient and Generative LLM-based Reinforcement Learning Agent for Recommenders | Arxiv 2024 | [Link] |
| CheatAgent: Attacking LLM-Empowered Recommender Systems via LLM Agent | KDD 2024 | [Link] |
| Laser: Parameter-Efficient LLM Bi-Tuning for Sequential Recommendation with Collaborative Information | Arxiv 2024 | [Link] |
| MARS: Matching Attribute-aware Representations for Text-based Sequential Recommendation | CIKM 2024 | [Link] |
| End-to-End Learnable Item Tokenization for Generative Recommendation | Arxiv 2024 | [Link] |
| Incorporate LLMs with Influential Recommender System | Arxiv 2024 | [Link] |
| Enhancing Sequential Recommendations through Multi-Perspective Reflections and Iteration | Arxiv 2024 | [Link] |
| STORE: Streamlining Semantic Tokenization and Generative Recommendation with A Single LLM | Arxiv 2024 | [Link] |
| Multilingual Prompts in LLM-Based Recommenders: Performance Across Languages | Arxiv 2024 | [Link] |
| Unleash LLMs Potential for Recommendation by Coordinating Twin-Tower Dynamic Semantic Token Generator | Arxiv 2024 | [Link] |
| Large Language Model Enhanced Hard Sample Identification for Denoising Recommendation | Arxiv 2024 | [Link] |
| Chain-of-thought prompting empowered generative user modeling for personalized recommendation | Neural Computing and Applications | [Link] |
| Challenging Fairness: A Comprehensive Exploration of Bias in LLM-Based Recommendations | Arxiv 2024 | [Link] |
| Decoding Style: Efficient Fine-Tuning of LLMs for Image-Guided Outfit Recommendation with Preference | CIKM 2024 | [Link] |
| LLM-Powered Text Simulation Attack Against ID-Free Recommender Systems | Arxiv 2024 | [Link] |
| FLARE: Fusing Language Models and Collaborative Architectures for Recommender Enhancement | Arxiv 2024 | [Link] |
| Retrieve, Annotate, Evaluate, Repeat: Leveraging Multimodal LLMs for Large-Scale Product Retrieval Evaluation | Arxiv 2024 | [Link] |
| HLLM: Enhancing Sequential Recommendations via Hierarchical Large Language Models for Item and User Modeling | Arxiv 2024 | [Link] |
| Large Language Model Ranker with Graph Reasoning for Zero-Shot Recommendation | ICANN 2024 | [Link] |
| User Knowledge Prompt for Sequential Recommendation | RecSys 2024 | [Link] |
| RLRF4Rec: Reinforcement Learning from Recsys Feedback for Enhanced Recommendation Reranking | Arxiv 2024 | [Link] |
| FELLAS: Enhancing Federated Sequential Recommendation with LLM as External Services | Arxiv 2024 | [Link] |
| TLRec: A Transfer Learning Framework to Enhance Large Language Models for Sequential Recommendation Tasks | RecSys 2024 | [Link] |
| SeCor: Aligning Semantic and Collaborative Representations by Large Language Models for Next-Point-of-Interest Recommendations | RecSys 2024 | [Link] |
| Efficient Inference for Large Language Model-based Generative Recommendation | Arxiv 2024 | [Link] |
| Instructing and Prompting Large Language Models for Explainable Cross-domain Recommendations | RecSys 2024 | [Link] |
| ReLand: Integrating Large Language Models' Insights into Industrial Recommenders via a Controllable Reasoning Pool | RecSys 2024 | [Link] |
| Inductive Generative Recommendation via Retrieval-based Speculation | Arxiv 2024 | [Link] |
| Constructing and Masking Preference Profile with LLMs for Filtering Discomforting Recommendation | Arxiv 2024 | [Link] |
| Towards Scalable Semantic Representation for Recommendation | Arxiv 2024 | [Link] |
| Large Language Models as Narrative-Driven Recommenders | Arxiv 2024 | [Link] |
| The Moral Case for Using Language Model Agents for Recommendation | Arxiv 2024 | [Link] |
| RosePO: Aligning LLM-based Recommenders with Human Values | Arxiv 2024 | [Link] |
| Comprehending Knowledge Graphs with Large Language Models for Recommender Systems | Arxiv 2024 | [Link] |
| Triple Modality Fusion: Aligning Visual, Textual, and Graph Data with Large Language Models for Multi-Behavior Recommendations | Arxiv 2024 | [Link] |
| Improving Pinterest Search Relevance Using Large Language Models | CIKM 2024 Workshop | [Link] |
| STAR: A Simple Training-free Approach for Recommendations using Large Language Models | Arxiv 2024 | [Link] |
| End-to-end Training for Recommendation with Language-based User Profiles | Arxiv 2024 | [Link] |
| Knowledge Graph Enhanced Language Agents for Recommendation | Arxiv 2024 | [Link] |
| Collaborative Knowledge Fusion: A Novel Approach for Multi-task Recommender Systems via LLMs | Arxiv 2024 | [Link] |
| Real-Time Personalization for LLM-based Recommendation with Customized In-Context Learning | Arxiv 2024 | [Link] |
| ReasoningRec: Bridging Personalized Recommendations and Human-Interpretable Explanations through LLM Reasoning | Arxiv 2024 | [Link] |
| Beyond Utility: Evaluating LLM as Recommender | Arxiv 2024 | [Link] |
| Enhancing ID-based Recommendation with Large Language Models | Arxiv 2024 | [Link] |
| LLM4PR: Improving Post-Ranking in Search Engine with Large Language Models | Arxiv 2024 | [Link] |
| Proactive Detection and Calibration of Seasonal Advertisements with Multimodal Large Language Models | Arxiv 2024 | [Link] |
| Enhancing ID-based Recommendation with Large Language Models | Arxiv 2024 | [Link] |
| Transferable Sequential Recommendation via Vector Quantized Meta Learning | Arxiv 2024 | [Link] |
| Self-Calibrated Listwise Reranking with Large Language Models | Arxiv 2024 | [Link] |
| Enhancing Large Language Model Based Sequential Recommender Systems with Pseudo Labels Reconstruction | ACL Findings 2024 | [Link] |
| Unleashing the Power of Large Language Models for Group POI Recommendations | Avrxi 2024 | [Link] |
| Scaling Laws for Online Advertisement Retrieval | Arxiv 2024 | [Link] |
| Explainable LLM-driven Multi-dimensional Distillation for E-Commerce Relevance Learning | Arxiv 2024 | [Link] |
| GOT4Rec: Graph of Thoughts for Sequential Recommendation | Arxiv 2024 | [Link] |
| HARec: Hyperbolic Graph-LLM Alignment for Exploration and Exploitation in Recommender Systems | Arxiv 2024 | [Link] |
| Cross-Domain Recommendation Meets Large Language Models | Arxiv 2024 | [Link] |
| Explainable CTR Prediction via LLM Reasoning | WSDM 2025 | [Link] |
| Enabling Explainable Recommendation in E-commerce with LLM-powered Product Knowledge Graph | IJCAI Workshop 2025 | [Link] |
| Break the ID-Language Barrier: An Adaption Framework for Sequential Recommendation | Arxiv 2024 | [Link] |
| LEADRE: Multi-Faceted Knowledge Enhanced LLM Empowered Display Advertisement Recommender System | Arxiv 2024 | [Link] |
| Pre-train, Align, and Disentangle: Empowering Sequential Recommendation with Large Language Models | Arxiv 2024 | [Link] |
| ULMRec: User-centric Large Language Model for Sequential Recommendation | Arxiv 2024 | [Link] |
| AltFS: Agency-light Feature Selection with Large Language Models in Deep Recommender Systems | Arxiv 2024 | [Link] |
| MRP-LLM: Multitask Reflective Large Language Models for Privacy-Preserving Next POI Recommendation | Arxiv 2024 | [Link] |
| MOPI-HFRS: A Multi-objective Personalized Health-aware Food Recommendation System with LLM-enhanced Interpretation | Arxiv 2024 | [Link] |
| SPRec: Leveraging Self-Play to Debias Preference Alignment for Large Language Model-based Recommendations | Arxiv 2024 | [Link] |
| RecSys Arena: Pair-wise Recommender System Evaluation with Large Language Models | Arxiv 2024 | [Link] |
| CRS Arena: Crowdsourced Benchmarking of Conversational Recommender Systems | Arxiv 2024 | [Link] |
| Boosting LLM-based Relevance Modeling with Distribution-Aware Robust Learning | Arxiv 2024 | [Link] |
| LLM is Knowledge Graph Reasoner: LLM's Intuition-aware Knowledge Graph Reasoning for Cold-start Sequential Recommendation | Arxiv 2024 | [Link] |
| Bridging the User-side Knowledge Gap in Knowledge-aware Recommendations with Large Language Models | Arxiv 2024 | [Link] |
| Sliding Windows Are Not the End: Exploring Full Ranking with Long-Context Large Language Models | Arxiv 2024 | [Link] |
| ChainRank-DPO: Chain Rank Direct Preference Optimization for LLM Rankers | Arxiv 2024 | [Link] |
| Are Longer Prompts Always Better? Prompt Selection in Large Language Models for Recommendation Systems | Arxiv 2024 | [Link] |
| Towards a Unified Paradigm: Integrating Recommendation Systems as a New Language in Large Models | Arxiv 2024 | [Link] |
| LLM-Powered User Simulator for Recommender System | Arxiv 2024 | [Link] |
| Enhancing Item Tokenization for Generative Recommendation through Self-Improvement | Arxiv 2024 | [Link] |
| Molar: Multimodal LLMs with Collaborative Filtering Alignment for Enhanced Sequential Recommendation | Arxiv 2024 | [Link] |
| An Automatic Graph Construction Framework based on Large Language Models for Recommendation | Arxiv 2024 | [Link] |
| RecLM: Recommendation Instruction Tuning | Arxiv 2024 | [Link] |
| The Efficiency vs. Accuracy Trade-off: Optimizing RAG-Enhanced LLM Recommender Systems Using Multi-Head Early Exit | Arxiv 2025 | [Link] |
| Knowledge Graph Retrieval-Augmented Generation for LLM-based Recommendation | Arxiv 2025 | [Link] |
| Efficient and Responsible Adaptation of Large Language Models for Robust and Equitable Top-k Recommendations | Avrxi 2025 | [Link] |
| Collaboration of Large Language Models and Small Recommendation Models for Device-Cloud Recommendation | KDD 2025 | [Link] |
| Guiding Retrieval using LLM-based Listwise Rankers | Avrxi 2025 | [Link] |
| Generative Retrieval for Book search | KDD 2025 | [Link] |
| Full-Stack Optimized Large Language Models for Lifelong Sequential Behavior Comprehension in Recommendation | Arxiv 2025 | [Link] |
| Large Language Model driven Policy Exploration for Recommender Systems | WSDM 2025 | [Link] |
| SampleLLM: Optimizing Tabular Data Synthesis in Recommendations | Arxiv 2025 | [Link] |
| PatchRec: Multi-Grained Patching for Efficient LLM-based Sequential Recommendation | Arxiv 2025 | [Link] |
| Uncertainty Quantification and Decomposition for LLM-based Recommendation | WWW 2025 | [Link] |
| A Zero-Shot Generalization Framework for LLM-Driven Cross-Domain Sequential Recommendation | Arxiv 2025 | [Link] |
| RankFlow: A Multi-Role Collaborative Reranking Workflow Utilizing Large Language Models | Arxiv 2025 | [Link] |
| FACTER: Fairness-Aware Conformal Thresholding and Prompt Engineering for Enabling Fair LLM-Based Recommender Systems | Arxiv 2025 | [Link] |
| Large Language Models Are Universal Recommendation Learners | Arxiv 2025 | [Link] |
| Intent Representation Learning with Large Language Model for Recommendation | Arxiv 2025 | [Link] |
| Boosting Knowledge Graph-based Recommendations through Confidence-Aware Augmentation with Large Language Models | Arxiv 2025 | [Link] |
| RALLRec: Improving Retrieval Augmented Large Language Model Recommendation with Representation Learning | WWW 2025 | [Link] |
| Solving the Content Gap in Roblox Game Recommendations: LLM-Based Profile Generation and Reranking | Arxiv 2025 | [Link] |
| MoLoRec: A Generalizable and Efficient Framework for LLM-Based Recommendation | Arxiv 2025 | [Link] |
| Unleashing the Power of Large Language Model for Denoising Recommendation | WWW 2025 | [Link] |
| Semantic Ads Retrieval at Walmart eCommerce with Language Models Progressively Trained on Multiple Knowledge Domains | Arxiv 2025 | [Link] |
| Order-agnostic Identifier for Large Language Model-based Generative Recommendation | Arxiv 2025 | [Link] |
| G-Refer: Graph Retrieval-Augmented Large Language Model for Explainable Recommendation | WWW 2025 | [Link] |
| LLM4Tag: Automatic Tagging System for Information Retrieval via Large Language Models | Arxiv 2025 | [Link] |
| Bursting Filter Bubble: Enhancing Serendipity Recommendations with Aligned Large Language Models | Arxiv 2025 | [Link] |
| ActionPiece: Contextually Tokenizing Action Sequences for Generative Recommendation | Arxiv 2025 | [Link] |
| TALKPLAY: Multimodal Music Recommendation with Large Language Models | Arxiv 2025 | [Link] |
| Enhancing Cross-Domain Recommendations with Memory-Optimized LLM-Based User Agents | Arxiv 2025 | [Link] |
| Enhancing LLM-Based Recommendations Through Personalized Reasoning | Arxiv 2025 | [Link] |
| Lost in Sequence: Do Large Language Models Understand Sequential Recommendation? | Arxiv 2025 | [Link] |
| InstructAgent: Building User Controllable Recommender via LLM Agent | Arxiv 2025 | [Link] |
| EAGER-LLM: Enhancing Large Language Models as Recommenders through Exogenous Behavior-Semantic Integration | Arxiv 2025 | [Link] |
| Efficient AI in Practice: Training and Deployment of Efficient LLMs for Industry Applications | Arxiv 2025 | [Link] |
| Collaborative Retrieval for Large Language Model-based Conversational Recommender Systems | Arxiv 2025 | [Link] |
| Active Large Language Model-based Knowledge Distillation for Session-based Recommendation | Arxiv 2025 | [Link] |
| Training Large Recommendation Models via Graph-Language Token Alignment | WWW 2025 | [Link] |
| PCL: Prompt-based Continual Learning for User Modeling in Recommender Systems | WWW 2025 | |
| FilterLLM: Text-To-Distribution LLM for Billion-Scale Cold-Start Recommendation | Arxiv 2025 | [Link] |
| Towards An Efficient LLM Training Paradigm for CTR Prediction | Arxiv 2025 | [Link] |
| LLMInit: A Free Lunch from Large Language Models for Selective Initialization of Recommendation | Arxiv 2025 | [Link] |
| PersonaX: A Recommendation Agent Oriented User Modeling Framework for Long Behavior Sequence | Arxiv 2025 | [Link] |
| Towards Next-Generation Recommender Systems: A Benchmark for Personalized Recommendation Assistant with LLMs | Arxiv 2025 | [Link] |
| Uncovering Cross-Domain Recommendation Ability of Large Language Models | Arxiv 2025 | [Link] |
| LLM-Driven Usefulness Labeling for IR Evaluation | Arxiv 2025 | [Link] |
| LREF: A Novel LLM-based Relevance Framework for E-commerce | WWW 2025 | [Link] |
| Process-Supervised LLM Recommenders via Flow-guided Tuning | Arxiv 2025 | [Link] |
| Image is All You Need: Towards Efficient and Effective Large Language Model-Based Recommender Systems | Arxiv 2025 | [Link] |
| Rank-R1: Enhancing Reasoning in LLM-based Document Rerankers via Reinforcement Learning | Arxiv 2025 | [Link] |
| Federated Cross-Domain Click-Through Rate Prediction With Large Language Model Augmentation | Arxiv 2025 | [Link] |
| BeLightRec: A lightweight recommender system enhanced with BERT | Arxiv 2025 | [Link] |
| RALLRec+: Retrieval Augmented Large Language Model Recommendation with Reasoning | Arxiv 2025 | [Link] |
| Alleviating LLM-based Generative Retrieval Hallucination in Alipay Search | Arxiv 2025 | [Link] |
| RuleAgent: Discovering Rules for Recommendation Denoising with Autonomous Language Agents | Arxiv 2025 | [Link] |
| CoRanking: Collaborative Ranking with Small and Large Ranking Agents | Arxiv 2025 | [Link] |
| Get the Agents Drunk: Memory Perturbations in Autonomous Agent-based Recommender Systems | Arxiv 2025 | [Link] |
| Rec-R1: Bridging Generative Large Language Models and User-Centric Recommendation Systems via Reinforcement Learning | Arxiv 2025 | [Link] |
| LLM-Augmented Graph Neural Recommenders: Integrating User Reviews | Arxiv 2025 | [Link] |
| Enhancing Embedding Representation Stability in Recommendation Systems with Semantic ID | Arxiv 2025 | [Link] |
| Retrieval-Augmented Purifier for Robust LLM-Empowered Recommendation | Arxiv 2025 | [Link] |
| Pre-training Generative Recommender with Multi-Identifier Item Tokenization | SIGIR 2025 | [Link] |
| LLM-Alignment Live-Streaming Recommendation | Arxiv 2025 | [Link] |
| Decoding Recommendation Behaviors of In-Context Learning LLMs Through Gradient Descent | Arxiv 2025 | [Link] |
| Automating Personalization: Prompt Optimization for Recommendation Reranking | Arxiv 2025 | [Link] |
| IterQR: An Iterative Framework for LLM-based Query Rewrite in e-Commercial Search System | Arxiv 2025 | [Link] |
| Multimodal Quantitative Language for Generative Recommendation | ICLR 2025 | [Link] |
| Coherency Improved Explainable Recommendation via Large Language Model | Arxiv 2025 | [Link] |
| VALUE: Value-Aware Large Language Model for Query Rewriting via Weighted Trie in Sponsored Search | Arxiv 2025 | [Link] |
| Large Language Models Enhanced Hyperbolic Space Recommender Systems | Arxiv 2025 | [Link] |
| Unified Generative Search and Recommendation | Arxiv 2025 | [Link] |
| Retrieval Augmented Generation with Collaborative Filtering for Personalized Text Generation | SIGIR 2025 | [Link] |
| StealthRank: LLM Ranking Manipulation via Stealthy Prompt Optimization | Arxiv 2025 | [Link] |
| PathGPT: Leveraging Large Language Models for Personalized Route Generation | Arxiv 2025 | [Link] |
| LLM4Ranking: An Easy-to-use Framework of Utilizing Large Language Models for Document Reranking | Arxiv 2025 | [Link] |
| How Good Are Large Language Models for Course Recommendation in MOOCs? | Avrxi 2025 | [Link] |
| Large Language Model Empowered Recommendation Meets All-domain Continual Pre-Training | Arxiv 2025 | [Link] |
| Enhancing LLM-based Recommendation through Semantic-Aligned Collaborative Knowledge | Arxiv 2025 | [Link] |
| HistLLM: A Unified Framework for LLM-Based Multimodal Recommendation with User History Encoding and Compression | Arxiv 2025 | [Link] |
| CROSSAN: Towards Efficient and Effective Adaptation of Multiple Multimodal Foundation Models for Sequential Recommendation | Arxiv 2025 | [Link] |
| PinRec: Outcome-Conditioned, Multi-Token Generative Retrieval for Industry-Scale Recommendation Systems | Arxiv 2025 | [Link] |
| Distilling Transitional Pattern to Large Language Models for Multimodal Session-based Recommendation | Arxiv 2025 | [Link] |
| Multi-Modal Hypergraph Enhanced LLM Learning for Recommendation | Arxiv 2025 | [Link] |
| Rethinking LLM-Based Recommendations: A Query Generation-Based, Training-Free Approach | Arxiv 2025 | [Link] |
| Generative Recommendation with Continuous-Token Diffusion | Arxiv 2025 | [Link] |
| From Reviews to Dialogues: Active Synthesis for Zero-Shot LLM-based Conversational Recommender System | Arxiv 2025 | [Link] |
| Killing Two Birds with One Stone: Unifying Retrieval and Ranking with a Single Generative Recommendation Model | SIGIR 2025 | [[Link(https://arxiv.org/abs/2504.16454) |
| Bridge the Domains: Large Language Models Enhanced Cross-domain Sequential Recommendation | SIGIR 2025 | [Link] |
| Search-Based Interaction For Conversation Recommendation via Generative Reward Model Based Simulated User | SIGIR 2025 | [Link] |
| Preserving Privacy and Utility in LLM-Based Product Recommendations | Arxiv 2025 | [Link] |
| Multi-agents based User Values Mining for Recommendation | Arxiv 2025 | [Link] |
| Who You Are Matters: Bridging Topics and Social Roles via LLM-Enhanced Logical Recommendation | Arxiv 2025 | [Link] |
| Explain What You Mean: Intent Augmented Knowledge Graph Recommender Built With LLM | Arxiv 2025 | [Link] |
| ThinkRec: Thinking-based recommendation via LLM | Arxiv 2025 | [Link] |
| DeepRec: Towards a Deep Dive Into the Item Space with Large Language Model Based Recommendation | Arxiv 2025 | [Link] |
| Bridging the Gap: Self-Optimized Fine-Tuning for LLM-based Recommender Systems | Arxiv 2025 | [Link] |
| What LLMs Miss in Recommendations: Bridging the Gap with Retrieval-Augmented Collaborative Signals | Arxiv 2025 | [Link] |
| Reinforced Latent Reasoning for LLM-based Recommendation | Arxiv 2025 | [Link] |
| AgentRecBench: Benchmarking LLM Agent-based Personalized Recommender Systems | Arxiv 2025 | [Link] |
| Reason-to-Recommend: Using Interaction-of-Thought Reasoning to Enhance LLM Recommendation | Arxiv 2025 | [Link] |
| Generating Long Semantic IDs in Parallel for Recommendation | KDD 2025 | [Link] |
| RecGPT: A Foundation Model for Sequential Recommendation | Arxiv 2025 | [Link] |
| Serendipitous Recommendation with Multimodal LLM | Arxiv 2025 | [Link] |
| Research on E-Commerce Long-Tail Product Recommendation Mechanism Based on Large-Scale Language Models | Arxiv 2025 | [Link] |
| LettinGo: Explore User Profile Generation for Recommendation System | Arxiv 2025 | [Link] |
| CORONA: A Coarse-to-Fine Framework for Graph-based Recommendation with Large Language Models | Arxiv 2025 | [Link] |
| CoVE: Compressed Vocabulary Expansion Makes Better LLM-based Recommender Systems | Arxiv 2025 | [Link] |
| LLM2Rec: Large Language Models Are Powerful Embedding Models for Sequential Recommendation | Arxiv 2025 | [Link] |
| Reinforcement Fine-Tuned Large Language Models for Next POI Recommendation | Arxiv 2025 | [Link] |
| IRanker: Towards Ranking Foundation Model | Arxiv 2025 | [Link] |
| LLM2Rec: Large Language Models Are Powerful Embedding Models for Sequential Recommendation | KDD 2025 | [Link] |
| FindRec: Stein-Guided Entropic Flow for Multi-Modal Sequential Recommendation | KDD 2025 | [Link] |
| Heterogeneous User Modeling for LLM-based Recommendation | RecSys 2025 | [Link] |
| BiFair: A Fairness-aware Training Framework for LLM-enhanced Recommender Systems via Bi-level Optimization | Arxiv 2025 | [Link] |
| CTR-Guided Generative Query Suggestion in Conversational Search | Arxiv 2025 | [Link] |
| Boosting Parameter Efficiency in LLM-Based Recommendation through Sophisticated Pruning | Arxiv 2025 | [Link] |
| A Language-Driven Framework for Improving Personalized Recommendations: Merging LLMs with Traditional Algorithms | Arxiv 2025 | [Link] |
| LLM-Driven Dual-Level Multi-Interest Modeling for Recommendation | Arxiv 2025 | [Link] |
| Revisiting Prompt Engineering: A Comprehensive Evaluation for LLM-based Personalized Recommendation | Arxiv 2025 | [Link] |
| R4ec: A Reasoning, Reflection, and Refinement Framework for Recommendation Systems | RecSys 2025 | [Link] |
| Exploring the Potential of LLMs for Serendipity Evaluation in Recommender Systems | RecSys 2025 | [Link] |
| Improving the Performance of Sequential Recommendation Systems with an Extended Large Language Model | Arxiv 2025 | [Link] |
| Integrating LLM-Derived Multi-Semantic Intent into Graph Model for Session-based Recommendation | Arxiv 2025 | [Link] |
| A Comprehensive Review on Harnessing Large Language Models to Overcome Recommender System Challenges | Arxiv 2025 | [Link] |
| End-to-End Personalization: Unifying Recommender Systems with Large Language Models | Arxiv 2025 | [Link] |
| Temporal User Profiling with LLMs: Balancing Short-Term and Long-Term Preferences for Recommendations | Arxiv 2025 | [Link] |
| LLM-Based Intelligent Agents for Music Recommendation: A Comparison with Classical Content-Based Filtering | Arxiv 2025 | [Link] |
| AdaptJobRec: Enhancing Conversational Career Recommendation through an LLM-Powered Agentic System | Arxiv 2025 | [Link] |
| LLM-Enhanced Linear Autoencoders for Recommendation | CIKM 2025 | [Link] |
| TrackRec: Iterative Alternating Feedback with Chain-of-Thought via Preference Alignment for Recommendation | Arxiv 2025 | [Link] |
| MMQ: Multimodal Mixture-of-Quantization Tokenization for Semantic ID Generation and User Behavioral Adaptation | Arxiv 2025 | [Link] |
| MLLMRec: Exploring the Potential of Multimodal Large Language Models in Recommender Systems | Arxiv 2025 | [Link] |
| Membership Inference Attacks on LLM-based Recommender Systems | Arxiv 2025 | [Link] |
| Revealing Potential Biases in LLM-Based Recommender Systems in the Cold Start Setting | Arxiv 2025 | [Link] |
| Efficient Item ID Generation for Large-Scale LLM-based Recommendation | Arxiv 2025 | [Link] |
| Knowledge-Augmented Relation Learning for Complementary Recommendation with Large Language Models | Arxiv 2025 | [Link] |
| Decoding in Latent Spaces for Efficient Inference in LLM-based Recommendation | EMNLP 2025 | [Link] |
| Learning Decomposed Contextual Token Representations from Pretrained and Collaborative Signals for Generative Recommendation | Arxiv 2025 | [Link] |
| LLM4Rec: Large Language Models for Multimodal Generative Recommendation with Causal Debiasing | Arxiv 2025 | [Link] |
| Empowering Denoising Sequential Recommendation with Large Language Model Embeddings | CIKM 2025 | [Link] |
| GRACE: Generative Representation Learning via Contrastive Policy Optimization | Arxiv 2025 | [Link] |
2. 数据集与基准测试
针对大语言模型相关推荐系统的数据集和基准测试,应保留原始的语义/文本特征,而非使用匿名化的特征ID。
2.1 数据集
| 数据集 | 推荐场景 | 链接 |
|---|---|---|
| RecSysLLMsP | 社交网络 | [链接] |
| AmazonQAC | 查询自动补全 | [链接] |
| NineRec | 9个领域 | [链接] |
| MicroLens | 视频流媒体 | [链接] |
| Amazon-Review 2023 | 电商 | [链接] |
| Reddit-Movie | 对话式 & 电影 | [链接] |
| Amazon-M2 | 电商 | [链接] |
| MovieLens | 电影 | [链接] |
| Amazon | 电商 | [链接] |
| BookCrossing | 书籍 | [链接] |
| GoodReads | 书籍 | [链接] |
| Anime | 动漫 | [链接] |
| PixelRec | 短视频 | [链接] |
| Netflix | 电影 | [链接] |
2.2 基准测试
| 基准测试 | 网页引用链接 | 论文 |
|---|---|---|
| InfoDeepSeek | [论文] | |
| RecBench | [论文] | |
| RecBench+ | [论文] | |
| Shopping MMLU | [论文] | |
| Amazon-M2 (KDD杯2023) | [链接] | [论文] |
| LLMRec | [链接] | [论文] |
| OpenP5 | [链接] | [论文] |
| TABLET | [链接] | [论文] |
3. 相关仓库
| 仓库名称 | 维护者 |
|---|---|
| rs-llm-paper-list | wwliu555 |
| awesome-recommend-system-pretraining-papers | archersama |
| LLM4Rec | WLiK |
| Awesome-LLM4RS-Papers | nancheng58 |
| LLM4IR-Survey | RUC-NLPIR |
贡献
👍 欢迎为本仓库贡献内容。
如果您发现了相关资源或本仓库中存在错误,请随时提交问题或拉取请求。
联系方式: chiangel [点] ljh [at] gmail [点] com
引用
@article{10.1145/3678004,
author = {Lin, Jianghao and Dai, Xinyi and Xi, Yunjia and Liu, Weiwen and Chen, Bo and Zhang, Hao and Liu, Yong and Wu, Chuhan and Li, Xiangyang and Zhu, Chenxu and Guo, Huifeng and Yu, Yong and Tang, Ruiming and Zhang, Weinan},
title = {推荐系统如何受益于大型语言模型:综述},
year = {2024},
publisher = {计算机协会},
address = {纽约,纽约州,美国},
issn = {1046-8188},
url = {https://doi.org/10.1145/3678004},
doi = {10.1145/3678004},
journal = {ACM信息系统汇刊},
month = {7月}
}
相似工具推荐
openclaw
OpenClaw 是一款专为个人打造的本地化 AI 助手,旨在让你在自己的设备上拥有完全可控的智能伙伴。它打破了传统 AI 助手局限于特定网页或应用的束缚,能够直接接入你日常使用的各类通讯渠道,包括微信、WhatsApp、Telegram、Discord、iMessage 等数十种平台。无论你在哪个聊天软件中发送消息,OpenClaw 都能即时响应,甚至支持在 macOS、iOS 和 Android 设备上进行语音交互,并提供实时的画布渲染功能供你操控。 这款工具主要解决了用户对数据隐私、响应速度以及“始终在线”体验的需求。通过将 AI 部署在本地,用户无需依赖云端服务即可享受快速、私密的智能辅助,真正实现了“你的数据,你做主”。其独特的技术亮点在于强大的网关架构,将控制平面与核心助手分离,确保跨平台通信的流畅性与扩展性。 OpenClaw 非常适合希望构建个性化工作流的技术爱好者、开发者,以及注重隐私保护且不愿被单一生态绑定的普通用户。只要具备基础的终端操作能力(支持 macOS、Linux 及 Windows WSL2),即可通过简单的命令行引导完成部署。如果你渴望拥有一个懂你
stable-diffusion-webui
stable-diffusion-webui 是一个基于 Gradio 构建的网页版操作界面,旨在让用户能够轻松地在本地运行和使用强大的 Stable Diffusion 图像生成模型。它解决了原始模型依赖命令行、操作门槛高且功能分散的痛点,将复杂的 AI 绘图流程整合进一个直观易用的图形化平台。 无论是希望快速上手的普通创作者、需要精细控制画面细节的设计师,还是想要深入探索模型潜力的开发者与研究人员,都能从中获益。其核心亮点在于极高的功能丰富度:不仅支持文生图、图生图、局部重绘(Inpainting)和外绘(Outpainting)等基础模式,还独创了注意力机制调整、提示词矩阵、负向提示词以及“高清修复”等高级功能。此外,它内置了 GFPGAN 和 CodeFormer 等人脸修复工具,支持多种神经网络放大算法,并允许用户通过插件系统无限扩展能力。即使是显存有限的设备,stable-diffusion-webui 也提供了相应的优化选项,让高质量的 AI 艺术创作变得触手可及。
everything-claude-code
everything-claude-code 是一套专为 AI 编程助手(如 Claude Code、Codex、Cursor 等)打造的高性能优化系统。它不仅仅是一组配置文件,而是一个经过长期实战打磨的完整框架,旨在解决 AI 代理在实际开发中面临的效率低下、记忆丢失、安全隐患及缺乏持续学习能力等核心痛点。 通过引入技能模块化、直觉增强、记忆持久化机制以及内置的安全扫描功能,everything-claude-code 能显著提升 AI 在复杂任务中的表现,帮助开发者构建更稳定、更智能的生产级 AI 代理。其独特的“研究优先”开发理念和针对 Token 消耗的优化策略,使得模型响应更快、成本更低,同时有效防御潜在的攻击向量。 这套工具特别适合软件开发者、AI 研究人员以及希望深度定制 AI 工作流的技术团队使用。无论您是在构建大型代码库,还是需要 AI 协助进行安全审计与自动化测试,everything-claude-code 都能提供强大的底层支持。作为一个曾荣获 Anthropic 黑客大奖的开源项目,它融合了多语言支持与丰富的实战钩子(hooks),让 AI 真正成长为懂上
ComfyUI
ComfyUI 是一款功能强大且高度模块化的视觉 AI 引擎,专为设计和执行复杂的 Stable Diffusion 图像生成流程而打造。它摒弃了传统的代码编写模式,采用直观的节点式流程图界面,让用户通过连接不同的功能模块即可构建个性化的生成管线。 这一设计巧妙解决了高级 AI 绘图工作流配置复杂、灵活性不足的痛点。用户无需具备编程背景,也能自由组合模型、调整参数并实时预览效果,轻松实现从基础文生图到多步骤高清修复等各类复杂任务。ComfyUI 拥有极佳的兼容性,不仅支持 Windows、macOS 和 Linux 全平台,还广泛适配 NVIDIA、AMD、Intel 及苹果 Silicon 等多种硬件架构,并率先支持 SDXL、Flux、SD3 等前沿模型。 无论是希望深入探索算法潜力的研究人员和开发者,还是追求极致创作自由度的设计师与资深 AI 绘画爱好者,ComfyUI 都能提供强大的支持。其独特的模块化架构允许社区不断扩展新功能,使其成为当前最灵活、生态最丰富的开源扩散模型工具之一,帮助用户将创意高效转化为现实。
markitdown
MarkItDown 是一款由微软 AutoGen 团队打造的轻量级 Python 工具,专为将各类文件高效转换为 Markdown 格式而设计。它支持 PDF、Word、Excel、PPT、图片(含 OCR)、音频(含语音转录)、HTML 乃至 YouTube 链接等多种格式的解析,能够精准提取文档中的标题、列表、表格和链接等关键结构信息。 在人工智能应用日益普及的今天,大语言模型(LLM)虽擅长处理文本,却难以直接读取复杂的二进制办公文档。MarkItDown 恰好解决了这一痛点,它将非结构化或半结构化的文件转化为模型“原生理解”且 Token 效率极高的 Markdown 格式,成为连接本地文件与 AI 分析 pipeline 的理想桥梁。此外,它还提供了 MCP(模型上下文协议)服务器,可无缝集成到 Claude Desktop 等 LLM 应用中。 这款工具特别适合开发者、数据科学家及 AI 研究人员使用,尤其是那些需要构建文档检索增强生成(RAG)系统、进行批量文本分析或希望让 AI 助手直接“阅读”本地文件的用户。虽然生成的内容也具备一定可读性,但其核心优势在于为机器
LLMs-from-scratch
LLMs-from-scratch 是一个基于 PyTorch 的开源教育项目,旨在引导用户从零开始一步步构建一个类似 ChatGPT 的大型语言模型(LLM)。它不仅是同名技术著作的官方代码库,更提供了一套完整的实践方案,涵盖模型开发、预训练及微调的全过程。 该项目主要解决了大模型领域“黑盒化”的学习痛点。许多开发者虽能调用现成模型,却难以深入理解其内部架构与训练机制。通过亲手编写每一行核心代码,用户能够透彻掌握 Transformer 架构、注意力机制等关键原理,从而真正理解大模型是如何“思考”的。此外,项目还包含了加载大型预训练权重进行微调的代码,帮助用户将理论知识延伸至实际应用。 LLMs-from-scratch 特别适合希望深入底层原理的 AI 开发者、研究人员以及计算机专业的学生。对于不满足于仅使用 API,而是渴望探究模型构建细节的技术人员而言,这是极佳的学习资源。其独特的技术亮点在于“循序渐进”的教学设计:将复杂的系统工程拆解为清晰的步骤,配合详细的图表与示例,让构建一个虽小但功能完备的大模型变得触手可及。无论你是想夯实理论基础,还是为未来研发更大规模的模型做准备