[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"similar-AlphaAvatar--AlphaAvatar":3,"tool-AlphaAvatar--AlphaAvatar":64},[4,17,27,35,43,56],{"id":5,"name":6,"github_repo":7,"description_zh":8,"stars":9,"difficulty_score":10,"last_commit_at":11,"category_tags":12,"status":16},3808,"stable-diffusion-webui","AUTOMATIC1111\u002Fstable-diffusion-webui","stable-diffusion-webui 是一个基于 Gradio 构建的网页版操作界面，旨在让用户能够轻松地在本地运行和使用强大的 Stable Diffusion 图像生成模型。它解决了原始模型依赖命令行、操作门槛高且功能分散的痛点，将复杂的 AI 绘图流程整合进一个直观易用的图形化平台。\n\n无论是希望快速上手的普通创作者、需要精细控制画面细节的设计师，还是想要深入探索模型潜力的开发者与研究人员，都能从中获益。其核心亮点在于极高的功能丰富度：不仅支持文生图、图生图、局部重绘（Inpainting）和外绘（Outpainting）等基础模式，还独创了注意力机制调整、提示词矩阵、负向提示词以及“高清修复”等高级功能。此外，它内置了 GFPGAN 和 CodeFormer 等人脸修复工具，支持多种神经网络放大算法，并允许用户通过插件系统无限扩展能力。即使是显存有限的设备，stable-diffusion-webui 也提供了相应的优化选项，让高质量的 AI 艺术创作变得触手可及。",162132,3,"2026-04-05T11:01:52",[13,14,15],"开发框架","图像","Agent","ready",{"id":18,"name":19,"github_repo":20,"description_zh":21,"stars":22,"difficulty_score":23,"last_commit_at":24,"category_tags":25,"status":16},1381,"everything-claude-code","affaan-m\u002Feverything-claude-code","everything-claude-code 是一套专为 AI 编程助手（如 Claude Code、Codex、Cursor 等）打造的高性能优化系统。它不仅仅是一组配置文件，而是一个经过长期实战打磨的完整框架，旨在解决 AI 代理在实际开发中面临的效率低下、记忆丢失、安全隐患及缺乏持续学习能力等核心痛点。\n\n通过引入技能模块化、直觉增强、记忆持久化机制以及内置的安全扫描功能，everything-claude-code 能显著提升 AI 在复杂任务中的表现，帮助开发者构建更稳定、更智能的生产级 AI 代理。其独特的“研究优先”开发理念和针对 Token 消耗的优化策略，使得模型响应更快、成本更低，同时有效防御潜在的攻击向量。\n\n这套工具特别适合软件开发者、AI 研究人员以及希望深度定制 AI 工作流的技术团队使用。无论您是在构建大型代码库，还是需要 AI 协助进行安全审计与自动化测试，everything-claude-code 都能提供强大的底层支持。作为一个曾荣获 Anthropic 黑客大奖的开源项目，它融合了多语言支持与丰富的实战钩子（hooks），让 AI 真正成长为懂上",138956,2,"2026-04-05T11:33:21",[13,15,26],"语言模型",{"id":28,"name":29,"github_repo":30,"description_zh":31,"stars":32,"difficulty_score":23,"last_commit_at":33,"category_tags":34,"status":16},2271,"ComfyUI","Comfy-Org\u002FComfyUI","ComfyUI 是一款功能强大且高度模块化的视觉 AI 引擎，专为设计和执行复杂的 Stable Diffusion 图像生成流程而打造。它摒弃了传统的代码编写模式，采用直观的节点式流程图界面，让用户通过连接不同的功能模块即可构建个性化的生成管线。\n\n这一设计巧妙解决了高级 AI 绘图工作流配置复杂、灵活性不足的痛点。用户无需具备编程背景，也能自由组合模型、调整参数并实时预览效果，轻松实现从基础文生图到多步骤高清修复等各类复杂任务。ComfyUI 拥有极佳的兼容性，不仅支持 Windows、macOS 和 Linux 全平台，还广泛适配 NVIDIA、AMD、Intel 及苹果 Silicon 等多种硬件架构，并率先支持 SDXL、Flux、SD3 等前沿模型。\n\n无论是希望深入探索算法潜力的研究人员和开发者，还是追求极致创作自由度的设计师与资深 AI 绘画爱好者，ComfyUI 都能提供强大的支持。其独特的模块化架构允许社区不断扩展新功能，使其成为当前最灵活、生态最丰富的开源扩散模型工具之一，帮助用户将创意高效转化为现实。",107662,"2026-04-03T11:11:01",[13,14,15],{"id":36,"name":37,"github_repo":38,"description_zh":39,"stars":40,"difficulty_score":23,"last_commit_at":41,"category_tags":42,"status":16},3704,"NextChat","ChatGPTNextWeb\u002FNextChat","NextChat 是一款轻量且极速的 AI 助手，旨在为用户提供流畅、跨平台的大模型交互体验。它完美解决了用户在多设备间切换时难以保持对话连续性，以及面对众多 AI 模型不知如何统一管理的痛点。无论是日常办公、学习辅助还是创意激发，NextChat 都能让用户随时随地通过网页、iOS、Android、Windows、MacOS 或 Linux 端无缝接入智能服务。\n\n这款工具非常适合普通用户、学生、职场人士以及需要私有化部署的企业团队使用。对于开发者而言，它也提供了便捷的自托管方案，支持一键部署到 Vercel 或 Zeabur 等平台。\n\nNextChat 的核心亮点在于其广泛的模型兼容性，原生支持 Claude、DeepSeek、GPT-4 及 Gemini Pro 等主流大模型，让用户在一个界面即可自由切换不同 AI 能力。此外，它还率先支持 MCP（Model Context Protocol）协议，增强了上下文处理能力。针对企业用户，NextChat 提供专业版解决方案，具备品牌定制、细粒度权限控制、内部知识库整合及安全审计等功能，满足公司对数据隐私和个性化管理的高标准要求。",87618,"2026-04-05T07:20:52",[13,26],{"id":44,"name":45,"github_repo":46,"description_zh":47,"stars":48,"difficulty_score":23,"last_commit_at":49,"category_tags":50,"status":16},2268,"ML-For-Beginners","microsoft\u002FML-For-Beginners","ML-For-Beginners 是由微软推出的一套系统化机器学习入门课程，旨在帮助零基础用户轻松掌握经典机器学习知识。这套课程将学习路径规划为 12 周，包含 26 节精炼课程和 52 道配套测验，内容涵盖从基础概念到实际应用的完整流程，有效解决了初学者面对庞大知识体系时无从下手、缺乏结构化指导的痛点。\n\n无论是希望转型的开发者、需要补充算法背景的研究人员，还是对人工智能充满好奇的普通爱好者，都能从中受益。课程不仅提供了清晰的理论讲解，还强调动手实践，让用户在循序渐进中建立扎实的技能基础。其独特的亮点在于强大的多语言支持，通过自动化机制提供了包括简体中文在内的 50 多种语言版本，极大地降低了全球不同背景用户的学习门槛。此外，项目采用开源协作模式，社区活跃且内容持续更新，确保学习者能获取前沿且准确的技术资讯。如果你正寻找一条清晰、友好且专业的机器学习入门之路，ML-For-Beginners 将是理想的起点。",84991,"2026-04-05T10:45:23",[14,51,52,53,15,54,26,13,55],"数据工具","视频","插件","其他","音频",{"id":57,"name":58,"github_repo":59,"description_zh":60,"stars":61,"difficulty_score":10,"last_commit_at":62,"category_tags":63,"status":16},3128,"ragflow","infiniflow\u002Fragflow","RAGFlow 是一款领先的开源检索增强生成（RAG）引擎，旨在为大语言模型构建更精准、可靠的上下文层。它巧妙地将前沿的 RAG 技术与智能体（Agent）能力相结合，不仅支持从各类文档中高效提取知识，还能让模型基于这些知识进行逻辑推理和任务执行。\n\n在大模型应用中，幻觉问题和知识滞后是常见痛点。RAGFlow 通过深度解析复杂文档结构（如表格、图表及混合排版），显著提升了信息检索的准确度，从而有效减少模型“胡编乱造”的现象，确保回答既有据可依又具备时效性。其内置的智能体机制更进一步，使系统不仅能回答问题，还能自主规划步骤解决复杂问题。\n\n这款工具特别适合开发者、企业技术团队以及 AI 研究人员使用。无论是希望快速搭建私有知识库问答系统，还是致力于探索大模型在垂直领域落地的创新者，都能从中受益。RAGFlow 提供了可视化的工作流编排界面和灵活的 API 接口，既降低了非算法背景用户的上手门槛，也满足了专业开发者对系统深度定制的需求。作为基于 Apache 2.0 协议开源的项目，它正成为连接通用大模型与行业专有知识之间的重要桥梁。",77062,"2026-04-04T04:44:48",[15,14,13,26,54],{"id":65,"github_repo":66,"name":67,"description_en":68,"description_zh":69,"ai_summary_zh":69,"readme_en":70,"readme_zh":71,"quickstart_zh":72,"use_case_zh":73,"hero_image_url":74,"owner_login":67,"owner_name":75,"owner_avatar_url":76,"owner_bio":77,"owner_company":78,"owner_location":78,"owner_email":79,"owner_twitter":78,"owner_website":78,"owner_url":80,"languages":81,"stars":98,"forks":99,"last_commit_at":100,"license":101,"difficulty_score":10,"env_os":102,"env_gpu":103,"env_ram":103,"env_deps":104,"category_tags":111,"github_topics":112,"view_count":23,"oss_zip_url":78,"oss_zip_packed_at":78,"status":16,"created_at":126,"updated_at":127,"faqs":128,"releases":129},1307,"AlphaAvatar\u002FAlphaAvatar","AlphaAvatar","A real-time interactive Omni Avatar built on LiveKit, which allows you to seamlessly integrate with any open source Avatar components (real-time model, visual, voice, memory, search, etc.).","AlphaAvatar 是一个可本地部署的「全能数字分身」框架，把实时 3D 形象、语音对话、长期记忆、任务规划、外部工具调用等功能打包成一套插件系统。它解决的是传统聊天机器人“用完即走、记不住事”的痛点，让 AI 真正成为长期陪伴、主动提醒、替你执行任务的私人管家。\n\n适合三类人：  \n• 开发者——想快速拼装属于自己的虚拟助手或游戏 NPC；  \n• 研究人员——需要可控环境做多模态记忆、自主规划实验；  \n• 注重隐私的普通用户——希望把健康、笔记、日程等数据留在本地，却仍享受智能助理的便利。\n\n亮点：完全开源、插件即插即用；支持 LiveKit 实时音视频；记忆模块跨文本、图像、声音统一存储；可调用 MCP、RAG、DeepResearch 等外部工具，实现“自己上网查资料、写总结、发邮件”。","\u003Cdiv align=\"center\"> \u003Ca name=\"readme-top\">\u003C\u002Fa>\n\n\u003Cpicture>\n  \u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"\u002F.github\u002Fbanner_dark.png\">\n  \u003Csource media=\"(prefers-color-scheme: light)\" srcset=\"\u002F.github\u002Fbanner_light.png\">\n  \u003Cimg style=\"width:100%;\" alt=\"The AvatarAlpha icon, the name of the repository.\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FAlphaAvatar_AlphaAvatar_readme_2471863ab522.png\">\n\u003C\u002Fpicture>\n\n\u003Cbr \u002F>\n\n[![PRs Welcome](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPRs-welcome!-brightgreen.svg?style=flat-square)](https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fpulls)\n[![GitHub last commit](https:\u002F\u002Fimg.shields.io\u002Fgithub\u002Flast-commit\u002FAlphaAvatar\u002FAlphaAvatar)](https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcommits\u002Fmain)\n[![License](https:\u002F\u002Fimg.shields.io\u002Fgithub\u002Flicense\u002FAlphaAvatar\u002FAlphaAvatar)](https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002FLICENSE)\n\n[![GitHub watchers](https:\u002F\u002Fimg.shields.io\u002Fgithub\u002Fwatchers\u002FAlphaAvatar\u002FAlphaAvatar?style=social&label=Watch)](https:\u002F\u002FGitHub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fwatchers\u002F?WT.mc_id=academic-105485-koreyst)\n[![GitHub forks](https:\u002F\u002Fimg.shields.io\u002Fgithub\u002Fforks\u002FAlphaAvatar\u002FAlphaAvatar?style=social&label=Fork)](https:\u002F\u002FGitHub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fnetwork\u002F?WT.mc_id=academic-105485-koreyst)\n[![GitHub stars](https:\u002F\u002Fimg.shields.io\u002Fgithub\u002Fstars\u002FAlphaAvatar\u002FAlphaAvatar?style=social&label=Star)](https:\u002F\u002FGitHub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fstargazers\u002F?WT.mc_id=academic-105485-koreyst)\n\n\n\u003Ch3 align=\"center\">\nLearnable, configurable, and pluggable Omni Personal Assistant for everyone\n\u003C\u002Fh3>\n\n\u003Cp align=\"center\">\n  \u003Ca href=\"ROADMAP.md\">ROADMAP\u003C\u002Fa>\n  |\n  \u003Ca href=\"#\">Demo\u003C\u002Fa>\n  |\n  \u003Ca href=\"#\">HomePage\u003C\u002Fa>\n  |\n  \u003Ca href=\"#\">Documents\u003C\u002Fa>\n\u003C\u002Fp>\n\n\u003C\u002Fdiv>\n\n---\n\n\u003Ch2>AlphaAvatar Introduction\u003C\u002Fh2>\n\nAlphaAvatar is an **Omni-Avatar personal assistant framework** designed to evolve into your **intelligent personal butler**.\n\n✨ **Fully self-hostable and privacy-first** — you can deploy AlphaAvatar locally or on your own infrastructure, with full control over your data, memory, and behavior.\n\nIt is built around a **plugin-based Agent architecture**, combining:\n\n- 🧠 Full-modality **Memory**\n- 🧬 Dynamic **Persona understanding**\n- 💡 Self-improving **Reflection**\n- 📅 Long-term **Planning & Execution**\n- 🧰 External **Tool integrations (MCP \u002F RAG \u002F DeepResearch)**\n- 😊 Real-time **Virtual Character (Avatar)**\n\nTogether, these components enable AlphaAvatar to move beyond a traditional chatbot into a **continuous, personalized, and proactive assistant system**.\n\n### 🎯 Vision\n\nOur goal is to build an AI that can **continuously learn, remember, and act on behalf of the user**, acting as a true **personal life manager** rather than a passive responder.\n\n### 🧩 What Can AlphaAvatar Do?\n\nWith the combination of internal Agent plugins and external tools, AlphaAvatar can:\n\n#### 1️⃣ Personal Data & Life Metrics Management\n- 📊 Track and analyze personal metrics (health, fitness, sleep, study progress)\n- 📈 Provide long-term insights and trend analysis\n- 🎯 Suggest improvements based on historical patterns\n\n#### 2️⃣ Knowledge & Notes Management\n- 📖 Organize personal notes, documents, and knowledge\n- 🔍 Retrieve relevant information via RAG\n- 🧠 Build a **personal knowledge base** over time\n\n#### 3️⃣ Task & Event Management\n- 📅 Schedule tasks and reminders\n- ⏰ Proactively notify based on context and priority\n- 🔄 Break down long-term goals into actionable steps\n\n#### 4️⃣ Autonomous Planning & Execution\n- 🧠 Plan multi-step workflows (learning plans, projects, research)\n- 🔧 Call tools automatically to complete tasks\n- 📌 Maintain consistency across long time horizons\n\n#### 5️⃣ Personalized Companion & Context Awareness\n- 🧬 Understand user preferences, habits, and personality\n- 💬 Provide highly personalized responses\n- 🤝 Maintain continuity across conversations and modalities\n\n#### 6️⃣ External World Interaction\n- 🌐 Search, research, and summarize real-world information\n- 🧰 Integrate with tools (email, database, APIs, messaging apps)\n- 🔗 Act as a bridge between user intent and external systems\n\n### 🤖 Why AlphaAvatar?\n\nUnlike traditional AI assistants, AlphaAvatar is:\n\n- **Stateful** → remembers and evolves with you\n- **Proactive** → acts, not just responds\n- **Composable** → fully plugin-driven architecture\n- **Omni-modal** → supports text, voice, and visual interaction\n- **Extensible** → easily integrates new tools and capabilities\n\n> 💡 AlphaAvatar is not just an assistant —\n> it is your **long-term AI companion and personal operating system**.\n\n---\n\n\u003Ch2>AlphaAvatar Plugins\u003C\u002Fh2>\n\n\u003Ctable>\n\u003Ctr>\n\u003Ctd width=\"50%\">\n\u003Ch3>🧠 Memory\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FIn_Progress-28a745?style=flat\" \u002F>\n\u003C\u002Fp>\n\u003Cp>Self-improving memory module for Omni-Avatar.\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-plugins\u002Favatar-plugins-memory\u002FREADME.md\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003Ctd width=\"50%\">\n\u003Ch3>🧬 Persona\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FIn_Progress-28a745?style=flat\" \u002F>\n\u003C\u002Fp>\n\u003Cp>Automatic extraction and real-time matching of user full modality persona.\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-plugins\u002Favatar-plugins-persona\u002FREADME.md\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003C\u002Ftr>\n\n\u003Ctr>\n\u003Ctd width=\"50%\">\n\u003Ch3>💡 Reflection\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPlanned-6c757d?style=flat\" alt=\"Planned\" \u002F>\n\u003C\u002Fp>\n\u003Cp>An Optimizer for Omni-Avatar that can automatically build an internal knowledge base for avatars.\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"#\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003Ctd width=\"50%\">\n\u003Ch3>📅 Planning\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPlanned-6c757d?style=flat\" alt=\"Planned\" \u002F>\n\u003C\u002Fp>\n\u003Cp>Agents need to plan over a longer time frame to ensure that their actions are sequential and reliable.\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"#\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003C\u002Ftr>\n\n\u003Ctr>\n\u003Ctd width=\"50%\">\n\u003Ch3>🤖 Behavior\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPlanned-6c757d?style=flat\" alt=\"Planned\" \u002F>\n\u003C\u002Fp>\n\u003Cp>Controls AlphaAvatar’s behavior logic and process flow.\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"#\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003Ctd width=\"50%\">\n\u003Ch3>😊 Virtual Character\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FIn_Progress-28a745?style=flat\" \u002F>\n\u003C\u002Fp>\n\u003Cp>The real-time generated \u003Cb>virtual character\u003C\u002Fb> that visually represents the Avatar during interactions.\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-plugins\u002Favatar-plugins-character\u002FREADME.md\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003C\u002Ftr>\n\u003C\u002Ftable>\n\n---\n\n\u003Ch2>Tools Plugins\u003C\u002Fh2>\n\n\u003Ctable>\n\u003Ctr>\n\u003Ctd width=\"50%\">\n\u003Ch3>🔍 DeepResearch\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FIn_Progress-28a745?style=flat\" \u002F>\n\u003C\u002Fp>\n\u003Cp>Allow AlphaAvatar to \u003Cstrong>access the network\u003C\u002Fstrong> and perform single-step\u002Fmulti-step inference through a separate Agent service to search for more accurate content.\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-plugins\u002Favatar-plugins-deepresearch\u002FREADME.md\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003Ctd width=\"50%\">\n\u003Ch3>📖 RAG\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FIn_Progress-28a745?style=flat\" \u002F>\n\u003C\u002Fp>\n\u003Cp>Allow AlphaAvatar to access \u003Cstrong>Documents\u002FSkills\u003C\u002Fstrong> (user-uploaded\u002Fgenerated by the Reflection module\u002FURL access) to obtain document-related information.\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-plugins\u002Favatar-plugins-rag\u002FREADME.md\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003C\u002Ftr>\n\n\u003Ctr>\n\u003Ctd width=\"50%\">\n\u003Ch3>🧰 MCP\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FIn_Progress-28a745?style=flat\" \u002F>\n\u003C\u002Fp>\n\u003Cp>Allows AlphaAvatar to \u003Cstrong>access real-world external tools\u003C\u002Fstrong>, such as databases, email, social media, etc.\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-plugins\u002Favatar-plugins-mcp\u002FREADME.md\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003Ctd width=\"50%\">\n\u003Ch3>🌍 SANDBOX\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPlanned-6c757d?style=flat\" alt=\"Planned\" \u002F>\n\u003C\u002Fp>\n\u003Cp>Provide AlphaAvatar with a sandbox environment to interact with the \u003Cstrong>external world or with other agents\u003C\u002Fstrong>, thereby enabling multi-agent interaction and exploration.\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"#\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003C\u002Ftr>\n\u003C\u002Ftable>\n\n---\n\n\u003Ch2>Docs and guides\u003C\u002Fh2>\n\n\u003Ch4>Latest News 🔥\u003C\u002Fh4>\n\n- [2026\u002F03] We have released AlphaAvatar **version 0.5.0** to support the MCP plugin, which enables retrieval and concurrent invocation of the MCP tools.\n  - Released AlphaAvatar **version 0.5.1**: Added [WhatsApp]((https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-channels\u002Favatar-channels-whatsapp\u002FREADME.md)) channel support via [Baileys](https:\u002F\u002Fgithub.com\u002Fwhiskeysockets\u002FBaileys) driver, enabling connection to AlphaAvatar Agent for **WhatsApp** integration.\n  - Released AlphaAvatar **version 0.5.2**: Add the AlphaAvatar Voice plugin to support the [Voice.ai](https:\u002F\u002Fvoice.ai\u002F) TTS API, providing a wider variety of speaker options.\n- [2026\u002F02] We have released AlphaAvatar **version 0.4.0** to support RAG by [RAG-Anything](https:\u002F\u002Fgithub.com\u002FHKUDS\u002FRAG-Anything) library and optimized the Memory and DeepResearch modules.\n  - Released AlphaAvatar **version 0.4.1**: Fix the Persona plugin bugs and Add new MCP plugin.\n- [2026\u002F01] We have released AlphaAvatar **version 0.3.0** to support DeepResearch by [tavily](https:\u002F\u002Ftavily.com) API.\n  - Released AlphaAvatar **version 0.3.1**: ADD tool calls during user-Assistant interactions to the Memory module.\n- [2025\u002F12] We have released AlphaAvatar **version 0.2.0** to support [AIRI](https:\u002F\u002Fgithub.com\u002Fmoeru-ai\u002Fairi) live2d-based virtual character display.\n- [2025\u002F11] We have released AlphaAvatar **version 0.1.0** to support automatic memory extraction, automatic user persona extraction and matching.\n\n\n\u003Cbr\u002F>\n\n\u003Ch2>Installation ⚙️\u003C\u002Fh2>\n\nInstall **stable** AlphaAvatar version from PyPI:\n\n```bash\nuv venv .my-env --python 3.11\nsource .my-env\u002Fbin\u002Factivate\npip install alpha-avatar-agents\n```\n\nInstall **latest** AlphaAvatar version from GitHub:\n\n```bash\ngit clone --recurse-submodules https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar.git\ncd AlphaAvatar\n\nuv venv .venv --python 3.11\nsource .venv\u002Fbin\u002Factivate\nuv sync --all-packages\n```\n\n\u003Ch2>Quick Start ⚡️\u003C\u002Fh2>\n\nStart your agent in dev mode to connect it to LiveKit and make it available from anywhere on the internet.\n\n---\n\n🧩 Step 1. Configure Environment Variables\n\n```bash\ncd AlphaAvatar\n\n# Copy template\ncp .env.template .env.dev\n```\n\nEdit .env.dev and set required environment variables.\n\n📦 Step 2. Download Required Files\n\n```bash\nalphaavatar download-files\n```\n\n✅ Step 3. Run the Agent\n\n```bash\nENV_FILE=.env.dev alphaavatar dev examples\u002Fagent_configs\u002Fpipeline_openai_airi.yaml\n# or\nENV_FILE=.env.dev alphaavatar dev examples\u002Fagent_configs\u002Fpipeline_openai_tools.yaml\n```\n\nTo see more supported modes, please refer to the [LiveKit doc](https:\u002F\u002Fdocs.livekit.io\u002Fagents\u002Fstart\u002Fvoice-ai\u002F).\n\nTo see more examples, please refer to the [Examples README](https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Fexamples\u002FREADME.md)\n\n\n\u003Ch2>Usage 🚀\u003C\u002Fh2>\n\nAlphaAvatar supports multiple **Access Channels**, allowing different types of users — from end users to developers — to interact with the system.\n\n---\n\n## 🧠 Runtime Architecture\n\n```\n            AlphaAvatar Runtime\n            ───────────────────\n\n    ┌──────────────────────────────┐\n    │        AgentSession          │\n    │        AvatarEngine          │\n    │   (LLM \u002F Memory \u002F RAG \u002F MCP) │\n    └──────────────┬───────────────┘\n                   │\n            InputDispatcher\n                   │\n            InputEnvelope\n                   │\n    ┌──────────────┴───────────────┐\n    │                              │\nChannel Adapters                Native Inputs\n(Ingress Layer)                (Web \u002F App)\n    │                              │\n    ▼                              ▼\nWhatsApp \u002F WeChat \u002F Slack      audio \u002F text \u002F video\n    │                              │\n    └──────────────┬───────────────┘\n                   ▼\n           OutputDispatcher\n                   │\n    ┌──────────────┴───────────────┐\n    │                              │\nChannel Egress                  Native Output\n(Messaging APIs)                (WebRTC \u002F UI)\n```\n\n> 💡 AlphaAvatar uses a **Channel Adapter architecture** to decouple runtime logic from communication channels.\n\n---\n\n\u003Ch3>🌐 Web Access\u003C\u002Fh3>\n\n\u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FIn_Development-orange?style=flat\" \u002F>\n\n🖥️ Browser-based interface for real-time interaction.\nThis will become the official AlphaAvatar user interface.\n\n- 🎙️ Real-time voice & multimodal communication\n- 🧠 Full plugin support (Memory \u002F RAG \u002F MCP \u002F etc.)\n- 😊 Virtual character display\n\n---\n\n\u003Ch3>💬 Social & Messaging Platforms\u003C\u002Fh3>\n\nInteract with AlphaAvatar directly inside messaging platforms.\n\nCapabilities:\n\n- 💬 Text-based conversation\n- 🎤 Voice message interaction\n- 🧰 Tool invocation via chat interface\n\n---\n\n\u003Ch4>WhatsApp\u003C\u002Fh4>\n\n\n\u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FAvailable-28a745?style=flat\" \u002F>\n\n📦 Channel introduction: [README](https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-channels\u002Favatar-channels-whatsapp\u002FREADME.md)\n\n▶️ Start WhatsApp Channel\n\n> Make sure AlphaAvatar Agent is already running (see Quick Start above).\n\n```bash\nENV_FILE=.env.dev sh examples\u002Fchannels\u002Fstart_whatsapp.sh\n````\n\n> 💡 The WhatsApp channel runs as an independent bridge process and connects to the Agent runtime.\n\n\u003Ch4>WeChat\u003C\u002Fh4>\n\n\u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPlanned-6c757d?style=flat\" \u002F>\n\n\u003Ch4>Slack\u003C\u002Fh4>\n\n\u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPlanned-6c757d?style=flat\" \u002F>\n\n---\n\n\u003Ch3>📲 Native Mobile App\u003C\u002Fh3>\n\n\u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPlanned-6c757d?style=flat\" \u002F>\n\nA dedicated AlphaAvatar mobile application providing:\n\n- 🎙️ Real-time voice communication\n- 😊 Live2D \u002F Virtual character visualization\n- 🧠 Persistent memory & persona\n\n---\n\n\u003Ch3>🧪 Developer Playground (Available Now)\u003C\u002Fh3>\n\n\u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FAvailable-28a745?style=flat\" \u002F>\n\nThis is the primary access channel for AlphaAvatar today.\n\nDevelopers can immediately access AlphaAvatar via the **LiveKit Playground**.\n\n👉 https:\u002F\u002Fagents-playground.livekit.io\u002F\n\nAfter starting your AlphaAvatar server:\n\n1. Connect to your LiveKit instance\n2. Configure the Agent name in the Playground (must match `avatar_name`, default: `Assistant`) to enable Explicit Dispatch.\n3. Connect to the agent room\n4. Start testing real-time interaction\n\nSupported capabilities:\n\n- 🎙️ Voice interaction\n- 🧠 Memory extraction\n- 🔍 RAG retrieval\n- 🧰 MCP tool invocation\n- 😊 Virtual character display\n\n![playground airi screenshot](.github\u002Fassets\u002Fplayground-airi-screenshot.png)\n\n---\n\n> 💡 AlphaAvatar is currently developer-first.\n> Web and mobile experiences are actively under development.\n","\u003Cdiv align=\"center\"> \u003Ca name=\"readme-top\">\u003C\u002Fa>\n\n\u003Cpicture>\n  \u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"\u002F.github\u002Fbanner_dark.png\">\n  \u003Csource media=\"(prefers-color-scheme: light)\" srcset=\"\u002F.github\u002Fbanner_light.png\">\n  \u003Cimg style=\"width:100%;\" alt=\"The AvatarAlpha icon, the name of the repository.\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FAlphaAvatar_AlphaAvatar_readme_2471863ab522.png\">\n\u003C\u002Fpicture>\n\n\u003Cbr \u002F>\n\n[![欢迎PR](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPRs-welcome!-brightgreen.svg?style=flat-square)](https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fpulls)\n[![GitHub最新提交](https:\u002F\u002Fimg.shields.io\u002Fgithub\u002Flast-commit\u002FAlphaAvatar\u002FAlphaAvatar)](https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcommits\u002Fmain)\n[![许可证](https:\u002F\u002Fimg.shields.io\u002Fgithub\u002Flicense\u002FAlphaAvatar\u002FAlphaAvatar)](https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002FLICENSE)\n\n[![GitHub关注者](https:\u002F\u002Fimg.shields.io\u002Fgithub\u002Fwatchers\u002FAlphaAvatar\u002FAlphaAvatar?style=social&label=Watch)](https:\u002F\u002FGitHub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fwatchers\u002F?WT.mc_id=academic-105485-koreyst)\n[![GitHub分支](https:\u002F\u002Fimg.shields.io\u002Fgithub\u002Fforks\u002FAlphaAvatar\u002FAlphaAvatar?style=social&label=Fork)](https:\u002F\u002FGitHub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fnetwork\u002F?WT.mc_id=academic-105485-koreyst)\n[![GitHub星标](https:\u002F\u002Fimg.shields.io\u002Fgithub\u002Fstars\u002FAlphaAvatar\u002FAlphaAvatar?style=social&label=Star)](https:\u002F\u002FGitHub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fstargazers\u002F?WT.mc_id=academic-105485-koreyst)\n\n\n\u003Ch3 align=\"center\">\n面向所有人的可学习、可配置、可插拔的全能个人助理\n\u003C\u002Fh3>\n\n\u003Cp align=\"center\">\n  \u003Ca href=\"ROADMAP.md\">路线图\u003C\u002Fa>\n  |\n  \u003Ca href=\"#\">演示\u003C\u002Fa>\n  |\n  \u003Ca href=\"#\">首页\u003C\u002Fa>\n  |\n  \u003Ca href=\"#\">文档\u003C\u002Fa>\n\u003C\u002Fp>\n\n\u003C\u002Fdiv>\n\n---\n\n\u003Ch2>AlphaAvatar简介\u003C\u002Fh2>\n\nAlphaAvatar是一个**全能型Avatar个人助理框架**，旨在逐步进化为你**智能的私人管家**。\n\n✨ **完全自托管且以隐私为先** —— 你可以将AlphaAvatar部署在本地或自己的基础设施上，全面掌控你的数据、记忆与行为。\n\n它基于**插件式Agent架构**构建，融合了以下核心能力：\n\n- 🧠 全模态**记忆**\n- 🧬 动态**人格理解**\n- 💡 自我提升的**反思**\n- 📅 长期**规划与执行**\n- 🧰 外部**工具集成（MCP \u002F RAG \u002F DeepResearch）**\n- 😊 实时**虚拟角色（Avatar）**\n\n这些组件共同使AlphaAvatar超越传统聊天机器人，成为一套**持续、个性化且主动的助理系统**。\n\n### 🎯 愿景\n\n我们的目标是打造一款能够**持续学习、记忆并代表用户采取行动**的AI，真正扮演**个人生活管理者**的角色，而非被动的响应者。\n\n### 🧩 AlphaAvatar能做什么？\n\n通过内部Agent插件与外部工具的结合，AlphaAvatar可以实现以下功能：\n\n#### 1️⃣ 个人数据与生活指标管理\n- 📊 跟踪并分析个人指标（健康、健身、睡眠、学习进度）\n- 📈 提供长期洞察与趋势分析\n- 🎯 根据历史模式提出改进建议\n\n#### 2️⃣ 知识与笔记管理\n- 📖 整理个人笔记、文档与知识\n- 🔍 通过RAG检索相关信息\n- 🧠 随着时间积累构建**个人知识库**\n\n#### 3️⃣ 任务与事件管理\n- 📅 安排任务与提醒\n- ⏰ 根据情境与优先级主动通知\n- 🔄 将长期目标分解为可执行步骤\n\n#### 4️⃣ 自主规划与执行\n- 🧠 规划多步骤工作流（学习计划、项目、研究）\n- 🔧 自动调用工具完成任务\n- 📌 在长时间跨度内保持一致性\n\n#### 5️⃣ 个性化陪伴与上下文感知\n- 🧬 理解用户的偏好、习惯与个性\n- 💬 提供高度个性化的回复\n- 🤝 在不同对话与模态间保持连续性\n\n#### 6️⃣ 对外部世界的交互\n- 🌐 搜索、研究并总结现实世界信息\n- 🧰 与各类工具集成（邮件、数据库、API、消息应用）\n- 🔗 作为连接用户意图与外部系统的桥梁\n\n### 🤖 为什么选择 AlphaAvatar？\n\n与传统 AI 助手不同，AlphaAvatar 具有以下特点：\n\n- **有状态** → 能够记住并与您共同成长\n- **主动** → 不仅响应，更能主动行动\n- **可组合** → 完全基于插件的架构\n- **多模态** → 支持文本、语音和视觉交互\n- **可扩展** → 轻松集成新工具与新能力\n\n> 💡 AlphaAvatar 不仅仅是一个助手——  \n> 它是您的**长期 AI 伙伴与个人操作系统**。\n\n---\n\n\u003Ch2>AlphaAvatar 插件\u003C\u002Fh2>\n\n\u003Ctable>\n\u003Ctr>\n\u003Ctd width=\"50%\">\n\u003Ch3>🧠 记忆\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FIn_Progress-28a745?style=flat\" \u002F>\n\u003C\u002Fp>\n\u003Cp>用于 Omni-Avatar 的自我提升记忆模块。\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-plugins\u002Favatar-plugins-memory\u002FREADME.md\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003Ctd width=\"50%\">\n\u003Ch3>🧬 人格\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FIn_Progress-28a745?style=flat\" \u002F>\n\u003C\u002Fp>\n\u003Cp>自动提取并实时匹配用户全模态人格。\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-plugins\u002Favatar-plugins-persona\u002FREADME.md\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003C\u002Ftr>\n\n\u003Ctr>\n\u003Ctd width=\"50%\">\n\u003Ch3>💡 反思\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPlanned-6c757d?style=flat\" alt=\"Planned\" \u002F>\n\u003C\u002Fp>\n\u003Cp>用于 Omni-Avatar 的优化器，能够自动构建 Avatar 的内部知识库。\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"#\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003Ctd width=\"50%\">\n\u003Ch3>📅 计划\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPlanned-6c757d?style=flat\" alt=\"Planned\" \u002F>\n\u003C\u002Fp>\n\u003Cp>Agent 需要进行更长时间跨度的规划，以确保其行为具有连贯性和可靠性。\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"#\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003C\u002Ftr>\n\n\u003Ctr>\n\u003Ctd width=\"50%\">\n\u003Ch3>🤖 行为\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPlanned-6c757d?style=flat\" alt=\"Planned\" \u002F>\n\u003C\u002Fp>\n\u003Cp>控制 AlphaAvatar 的行为逻辑与流程。\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"#\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003Ctd width=\"50%\">\n\u003Ch3>😊 虚拟角色\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FIn_Progress-28a745?style=flat\" \u002F>\n\u003C\u002Fp>\n\u003Cp>在交互过程中实时生成的、用于视觉呈现 Avatar 的\u003Cb>虚拟角色\u003C\u002Fb>。\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-plugins\u002Favatar-plugins-character\u002FREADME.md\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003C\u002Ftr>\n\u003C\u002Ftable>\n\n---\n\n\u003Ch2>工具插件\u003C\u002Fh2>\n\n\u003Ctable>\n\u003Ctr>\n\u003Ctd width=\"50%\">\n\u003Ch3>🔍 深度研究\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FIn_Progress-28a745?style=flat\" \u002F>\n\u003C\u002Fp>\n\u003Cp>允许 AlphaAvatar \u003Cstrong>访问网络\u003C\u002Fstrong>，并通过独立的 Agent 服务执行单步或多步推理，以搜索更精准的内容。\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-plugins\u002Favatar-plugins-deepresearch\u002FREADME.md\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003Ctd width=\"50%\">\n\u003Ch3>📖 RAG\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FIn_Progress-28a745?style=flat\" \u002F>\n\u003C\u002Fp>\n\u003Cp>允许 AlphaAvatar 访问 \u003Cstrong>文档\u002F技能\u003C\u002Fstrong>（用户上传、由反思模块生成或通过 URL 访问），以获取与文档相关的信息。\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-plugins\u002Favatar-plugins-rag\u002FREADME.md\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003C\u002Ftr>\n\n\u003Ctr>\n\u003Ctd width=\"50%\">\n\u003Ch3>🧰 MCP\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FIn_Progress-28a745?style=flat\" \u002F>\n\u003C\u002Fp>\n\u003Cp>允许 AlphaAvatar \u003Cstrong>访问现实世界的外部工具\u003C\u002Fstrong>,例如数据库、电子邮件、社交媒体等。\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-plugins\u002Favatar-plugins-mcp\u002FREADME.md\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003Ctd width=\"50%\">\n\u003Ch3>🌍 沙盒\u003C\u002Fh3>\n\u003Cp>\n  \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPlanned-6c757d?style=flat\" alt=\"Planned\" \u002F>\n\u003C\u002Fp>\n\u003Cp>为 AlphaAvatar 提供沙盒环境，使其能够与\u003Cstrong>外部世界或其他 agent\u003C\u002Fstrong>互动，从而实现多 agent 交互与探索。\u003C\u002Fp>\n\u003Cp>\n\u003Ca href=\"#\">README↗\u003C\u002Fa>\n\u003C\u002Fp>\n\u003C\u002Ftd>\n\u003C\u002Ftr>\n\u003C\u002Ftable>\n\n---\n\n\u003Ch2>文档与指南\u003C\u002Fh2>\n\n\u003Ch4>最新动态 🔥\u003C\u002Fh4>\n\n- [2026\u002F03] 我们发布了 AlphaAvatar **版本 0.5.0**，支持 MCP 插件，该插件可实现对 MCP 工具的检索与并发调用。\n  - 发布了 AlphaAvatar **版本 0.5.1**：新增了 [WhatsApp]((https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-channels\u002Favatar-channels-whatsapp\u002FREADME.md)) 渠道支持，通过 [Baileys](https:\u002F\u002Fgithub.com\u002Fwhiskeysockets\u002FBaileys) 驱动程序连接到 AlphaAvatar Agent，实现 **WhatsApp** 集成。\n  - 发布了 AlphaAvatar **版本 0.5.2**：新增 AlphaAvatar Voice 插件，支持 [Voice.ai](https:\u002F\u002Fvoice.ai\u002F) TTS API，提供更丰富的说话人选项。\n- [2026\u002F02] 我们发布了 AlphaAvatar **版本 0.4.0**，支持由 [RAG-Anything](https:\u002F\u002Fgithub.com\u002FHKUDS\u002FRAG-Anything) 库提供的 RAG，并优化了记忆与深度研究模块。\n  - 发布了 AlphaAvatar **版本 0.4.1**：修复了人格插件的 bug，并新增了 MCP 插件。\n- [2026\u002F01] 我们发布了 AlphaAvatar **版本 0.3.0**，支持由 [tavily](https:\u002F\u002Ftavily.com) API 提供的深度研究。\n  - 发布了 AlphaAvatar **版本 0.3.1**：在记忆模块中新增了用户与助手交互时的工具调用功能。\n- [2025\u002F12] 我们发布了 AlphaAvatar **版本 0.2.0**，支持基于 [AIRI](https:\u002F\u002Fgithub.com\u002Fmoeru-ai\u002Fairi) live2d 的虚拟角色展示。\n- [2025\u002F11] 我们发布了 AlphaAvatar **版本 0.1.0**，支持自动记忆提取、自动用户人格提取与匹配。\n\n\n\u003Cbr\u002F>\n\n\u003Ch2>安装 ⚙️\u003C\u002Fh2>\n\n从 PyPI 安装 **稳定版** AlphaAvatar：\n\n```bash\nuv venv .my-env --python 3.11\nsource .my-env\u002Fbin\u002Factivate\npip install alpha-avatar-agents\n```\n\n从 GitHub 安装 **最新版** AlphaAvatar：\n\n```bash\ngit clone --recurse-submodules https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar.git\ncd AlphaAvatar\n\nuv venv .venv --python 3.11\nsource .venv\u002Fbin\u002Factivate\nuv sync --all-packages\n```\n\n\u003Ch2>快速入门 ⚡️\u003C\u002Fh2>\n\n以开发模式启动您的 agent，将其连接到 LiveKit，并使其可在互联网上的任何地方使用。\n\n---\n\n🧩 第一步：配置环境变量\n\n```bash\ncd AlphaAvatar\n\n# 复制模板\ncp .env.template .env.dev\n```\n\n编辑 .env.dev 并设置所需的环境变量。\n\n📦 第二步：下载所需文件\n\n```bash\nalphaavatar download-files\n```\n\n✅ 第三步：运行 agent\n\n```bash\nENV_FILE=.env.dev alphaavatar dev examples\u002Fagent_configs\u002Fpipeline_openai_airi.yaml\n\n# 或\nENV_FILE=.env.dev alphaavatar dev examples\u002Fagent_configs\u002Fpipeline_openai_tools.yaml\n```\n\n要查看更多支持的模式，请参阅 [LiveKit 文档](https:\u002F\u002Fdocs.livekit.io\u002Fagents\u002Fstart\u002Fvoice-ai\u002F)。\n\n要查看更多示例，请参阅 [示例自述文件](https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Fexamples\u002FREADME.md)\n\n\n\u003Ch2>使用方法 🚀\u003C\u002Fh2>\n\nAlphaAvatar 支持多种**接入渠道**，使不同类型的用户——从终端用户到开发者——都能与系统进行交互。\n\n---\n\n## 🧠 运行时架构\n\n```\n            AlphaAvatar 运行时\n            ───────────────────\n\n    ┌──────────────────────────────┐\n    │        代理会话          │\n    │        虚拟形象引擎          │\n    │   (LLM \u002F 内存 \u002F RAG \u002F MCP) │\n    └──────────────┬───────────────┘\n                   │\n            输入分发器\n                   │\n            输入信封\n                   │\n    ┌──────────────┴───────────────┐\n    │                              │\n通道适配器                原生输入\n（入口层）                （Web \u002F 应用）\n    │                              │\n    ▼                              ▼\nWhatsApp \u002F 微信 \u002F Slack      音频 \u002F 文本 \u002F 视频\n    │                              │\n    └──────────────┬───────────────┘\n                   ▼\n           输出分发器\n                   │\n    ┌──────────────┴───────────────┐\n    │                              │\n通道出口                  原生输出\n（消息API）                （WebRTC \u002F 界面）\n```\n\n> 💡 AlphaAvatar 采用**通道适配器架构**，将运行时逻辑与通信渠道解耦。\n\n---\n\n\u003Ch3>🌐 Web 访问\u003C\u002Fh3>\n\n\u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FIn_Development-orange?style=flat\" \u002F>\n\n🖥️ 基于浏览器的实时交互界面。\n这将成为 AlphaAvatar 的官方用户界面。\n\n- 🎙️ 实时语音与多模态通信\n- 🧠 完整插件支持（内存 \u002F RAG \u002F MCP 等）\n- 😊 虚拟角色展示\n\n---\n\n\u003Ch3>💬 社交与消息平台\u003C\u002Fh3>\n\n直接在消息平台上与 AlphaAvatar 互动。\n\n功能：\n\n- 💬 文本对话\n- 🎤 语音消息交互\n- 🧰 通过聊天界面调用工具\n\n---\n\n\u003Ch4>WhatsApp\u003C\u002Fh4>\n\n\n\u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FAvailable-28a745?style=flat\" \u002F>\n\n📦 渠道介绍：[自述文件](https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fblob\u002Fmain\u002Favatar-channels\u002Favatar-channels-whatsapp\u002FREADME.md)\n\n▶️ 启动 WhatsApp 渠道\n\n> 确保 AlphaAvatar 代理已运行（参见上方的快速入门）。\n\n```bash\nENV_FILE=.env.dev sh examples\u002Fchannels\u002Fstart_whatsapp.sh\n````\n\n> 💡 WhatsApp 渠道以独立桥接进程运行，并连接到代理运行时。\n\n\u003Ch4>微信\u003C\u002Fh4>\n\n\u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPlanned-6c757d?style=flat\" \u002F>\n\n\u003Ch4>Slack\u003C\u002Fh4>\n\n\u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPlanned-6c757d?style=flat\" \u002F>\n\n---\n\n\u003Ch3>📲 原生移动应用\u003C\u002Fh3>\n\n\u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FPlanned-6c757d?style=flat\" \u002F>\n\n一款专门的 AlphaAvatar 移动应用程序，提供：\n\n- 🎙️ 实时语音通信\n- 😊 Live2D \u002F 虚拟角色可视化\n- 🧠 持久化内存与人格设定\n\n---\n\n\u003Ch3>🧪 开发者游乐场（现已可用）\u003C\u002Fh3>\n\n\u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FAvailable-28a745?style=flat\" \u002F>\n\n这是目前 AlphaAvatar 的主要接入渠道。\n\n开发者可通过 **LiveKit 游乐场** 立即访问 AlphaAvatar。\n\n👉 https:\u002F\u002Fagents-playground.livekit.io\u002F\n\n启动 AlphaAvatar 服务器后：\n\n1. 连接到您的 LiveKit 实例\n2. 在游乐场中配置代理名称（必须与 `avatar_name` 匹配，默认为 `Assistant`），以启用显式分发。\n3. 连接到代理房间\n4. 开始测试实时交互\n\n支持的功能：\n\n- 🎙️ 语音交互\n- 🧠 内存提取\n- 🔍 RAG 检索\n- 🧩 MCP 工具调用\n- 😊 虚拟角色展示\n\n![playground airi 截图](.github\u002Fassets\u002Fplayground-airi-screenshot.png)\n\n---\n\n> 💡 AlphaAvatar 目前以开发者为先。\n> Web 和移动体验正在积极开发中。","# AlphaAvatar 快速上手指南（中文版）\n\n## 环境准备\n- **操作系统**：Linux \u002F macOS \u002F Windows WSL2  \n- **Python**：3.11（必须）  \n- **工具链**：  \n  - `uv`（推荐，安装：`curl -Ls https:\u002F\u002Fastral.sh\u002Fuv\u002Finstall.sh | bash`）  \n  - Git（含子模块支持）\n\n## 安装步骤\n\n### 1. 稳定版（PyPI）\n```bash\nuv venv .my-env --python 3.11\nsource .my-env\u002Fbin\u002Factivate\npip install alpha-avatar-agents\n```\n\n### 2. 最新版（GitHub）\n```bash\ngit clone --recurse-submodules https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar.git\ncd AlphaAvatar\n\nuv venv .venv --python 3.11\nsource .venv\u002Fbin\u002Factivate\nuv sync --all-packages\n```\n\n## 基本使用\n\n### 1. 配置环境变量\n```bash\ncd AlphaAvatar\ncp .env.template .env.dev\n# 编辑 .env.dev，填入必要配置（如 OpenAI Key、LiveKit 等）\n```\n\n### 2. 下载依赖文件\n```bash\nalphaavatar download-files\n```\n\n### 3. 启动开发模式\n```bash\nENV_FILE=.env.dev alphaavatar dev examples\u002Fagent_configs\u002Fpipeline_openai_airi.yaml\n# 或\nENV_FILE=.env.dev alphaavatar dev examples\u002Fagent_configs\u002Fpipeline_openai_tools.yaml\n```\n\n启动后，AlphaAvatar 将通过 LiveKit 暴露公网访问地址，即可在浏览器或 WhatsApp\u002F微信\u002FSlack 等渠道与它对话。","**背景**：独立开发者林可正在做一款面向自由职业者的「时间-健康-收入」一体化管理 SaaS，需要快速上线一个 24×7 的 AI 健康教练，既能实时语音互动，又能记住用户长期数据并给出个性化建议。\n\n### 没有 AlphaAvatar 时\n- **数据孤岛**：健康、日程、收入分别存在 Notion、Google Fit、Stripe，每次查询都要手动切换，无法交叉分析。\n- **实时性差**：用传统 WebSocket 做语音对话，延迟 2-3 秒，用户说“我今天跑了 5 公里”后，系统 5 秒才回“很好”，体验尴尬。\n- **记忆断档**：用户上周提到“膝盖旧伤”，今天再问训练计划，系统完全忘了，只能重复提问。\n- **开发周期长**：为了集成 3D Avatar、语音合成、RAG 检索，需要分别对接 LiveKit、ElevenLabs、LlamaIndex，排期至少 6 周。\n- **隐私顾虑**：用户担心健康数据上云，团队被迫自建加密存储，额外增加 2 人月运维成本。\n\n### 使用 AlphaAvatar 后\n- **一站式数据管道**：AlphaAvatar 的 MCP 插件把 Notion、Google Fit、Stripe 拉通，自动聚合为「健康-时间-收入」统一视图，实时生成洞察。\n- **毫秒级语音互动**：基于 LiveKit 的 Omni Avatar，端到端延迟 \u003C300 ms，用户边跑边说“心率 160 正常吗？”Avatar 立即语音回复并同步弹出 3D 表情鼓励。\n- **长期记忆连续**：AlphaAvatar 把“膝盖旧伤”写进本地向量记忆，后续每次生成训练计划都自动降低跳跃类动作权重，用户直呼“比私教还懂我”。\n- **3 天上线**：直接复用 AlphaAvatar 的插件市场，勾选「健康教练」模板，再替换自有 3D 模型，核心功能 72 小时跑通，省下 5 周开发时间。\n- **完全本地部署**：Docker 一键拉起，所有数据留在用户 NAS，GDPR 合规审计一次通过，运维成本归零。\n\n**一句话总结**：AlphaAvatar 让独立团队在 3 天内交付一个会听、会记、会思考、还能本地部署的 24×7 健康教练，把 6 周开发周期压缩到 72 小时，同时把用户隐私牢牢锁在本地。","https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FAlphaAvatar_AlphaAvatar_79235bab.png","Licheng Wang","https:\u002F\u002Foss.gittoolsai.com\u002Favatars\u002FAlphaAvatar_1de009b5.png","Ultimate efficiency and elegance.",null,"244267620@qq.com","https:\u002F\u002Fgithub.com\u002FAlphaAvatar",[82,86,90,94],{"name":83,"color":84,"percentage":85},"Python","#3572A5",97.2,{"name":87,"color":88,"percentage":89},"TypeScript","#3178c6",1.4,{"name":91,"color":92,"percentage":93},"Shell","#89e051",1.1,{"name":95,"color":96,"percentage":97},"Makefile","#427819",0.3,566,22,"2026-04-02T09:59:22","Apache-2.0","Linux, macOS","未说明",{"notes":105,"python":106,"dependencies":107},"需先安装 uv 并创建 Python 3.11 虚拟环境；首次运行需执行 alphaavatar download-files 下载所需文件；支持通过 LiveKit 提供公网语音\u002F视频接入；支持 WhatsApp、WeChat、Slack 等多通道接入；插件体系仍在快速迭代，部分功能处于 In Progress 状态","3.11",[108,109,110],"alpha-avatar-agents","uv","livekit",[53,14,26,55,13,15,54],[113,114,110,115,116,117,118,119,120,121,122,123,124,125],"agent","avatar","llm","memory","omni","plugin","real-time","search","speech","visual","ai","virtual-character","reflection","2026-03-27T02:49:30.150509","2026-04-06T05:16:37.569054",[],[130,135,140,145,150,155,160,165,170],{"id":131,"version":132,"summary_zh":133,"released_at":134},105595,"v0.5.2","\u003C!-- Release notes generated using configuration in .github\u002Frelease.yml at v0.5.2 -->\n\n\n\n**Full Changelog**: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.5.1...v0.5.2\n\nFull Changelog: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.5.1...v0.5.2\n","2026-03-29T19:18:38",{"id":136,"version":137,"summary_zh":138,"released_at":139},105596,"v0.5.1","\u003C!-- Release notes generated using configuration in .github\u002Frelease.yml at v0.5.1 -->\n\n\n\n**Full Changelog**: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.5.0...v0.5.1\n\nFull Changelog: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.5.0...v0.5.1\n","2026-03-21T11:25:46",{"id":141,"version":142,"summary_zh":143,"released_at":144},105597,"v0.5.0","\u003C!-- Release notes generated using configuration in .github\u002Frelease.yml at v0.5.0 -->\n\n\n\n**Full Changelog**: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.4.1...v0.5.0\n\nFull Changelog: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.4.1...v0.5.0\n","2026-02-21T11:24:12",{"id":146,"version":147,"summary_zh":148,"released_at":149},105598,"v0.4.1","\u003C!-- Release notes generated using configuration in .github\u002Frelease.yml at v0.4.1 -->\n\n\n\n**Full Changelog**: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.4.0...v0.4.1\n\nFull Changelog: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.4.0...v0.4.1\n","2026-02-08T10:19:22",{"id":151,"version":152,"summary_zh":153,"released_at":154},105599,"v0.4.0","\u003C!-- Release notes generated using configuration in .github\u002Frelease.yml at v0.4.0 -->\n\n\n\n**Full Changelog**: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.3.1...v0.4.0\n\nFull Changelog: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.3.1...v0.4.0\n","2026-02-01T11:25:38",{"id":156,"version":157,"summary_zh":158,"released_at":159},105600,"v0.3.1","\u003C!-- Release notes generated using configuration in .github\u002Frelease.yml at v0.3.1 -->\n\n\n\n**Full Changelog**: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.3.0...v0.3.1\n\nFull Changelog: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.3.0...v0.3.1\n","2026-01-04T11:20:53",{"id":161,"version":162,"summary_zh":163,"released_at":164},105601,"v0.3.0","\u003C!-- Release notes generated using configuration in .github\u002Frelease.yml at v0.3.0 -->\n\n\n\n**Full Changelog**: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.2.0...v0.3.0\n\nFull Changelog: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.2.0...v0.3.0\n","2025-12-31T12:51:02",{"id":166,"version":167,"summary_zh":168,"released_at":169},105602,"v0.2.0","\u003C!-- Release notes generated using configuration in .github\u002Frelease.yml at v0.2.0 -->\n\n\n\n**Full Changelog**: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.1.0...v0.2.0\n\nFull Changelog: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcompare\u002Fv0.1.0...v0.2.0\n","2025-12-06T11:01:41",{"id":171,"version":172,"summary_zh":173,"released_at":174},105603,"v0.1.0","\u003C!-- Release notes generated using configuration in .github\u002Frelease.yml at v0.1.0 -->\n\n\n\n**Full Changelog**: https:\u002F\u002Fgithub.com\u002FAlphaAvatar\u002FAlphaAvatar\u002Fcommits\u002Fv0.1.0\n","2025-11-13T06:12:37"]