[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"similar-google-deepmind--dm_control":3,"tool-google-deepmind--dm_control":64},[4,17,27,35,43,56],{"id":5,"name":6,"github_repo":7,"description_zh":8,"stars":9,"difficulty_score":10,"last_commit_at":11,"category_tags":12,"status":16},3808,"stable-diffusion-webui","AUTOMATIC1111\u002Fstable-diffusion-webui","stable-diffusion-webui 是一个基于 Gradio 构建的网页版操作界面，旨在让用户能够轻松地在本地运行和使用强大的 Stable Diffusion 图像生成模型。它解决了原始模型依赖命令行、操作门槛高且功能分散的痛点，将复杂的 AI 绘图流程整合进一个直观易用的图形化平台。\n\n无论是希望快速上手的普通创作者、需要精细控制画面细节的设计师，还是想要深入探索模型潜力的开发者与研究人员，都能从中获益。其核心亮点在于极高的功能丰富度：不仅支持文生图、图生图、局部重绘（Inpainting）和外绘（Outpainting）等基础模式，还独创了注意力机制调整、提示词矩阵、负向提示词以及“高清修复”等高级功能。此外，它内置了 GFPGAN 和 CodeFormer 等人脸修复工具，支持多种神经网络放大算法，并允许用户通过插件系统无限扩展能力。即使是显存有限的设备，stable-diffusion-webui 也提供了相应的优化选项，让高质量的 AI 艺术创作变得触手可及。",162132,3,"2026-04-05T11:01:52",[13,14,15],"开发框架","图像","Agent","ready",{"id":18,"name":19,"github_repo":20,"description_zh":21,"stars":22,"difficulty_score":23,"last_commit_at":24,"category_tags":25,"status":16},1381,"everything-claude-code","affaan-m\u002Feverything-claude-code","everything-claude-code 是一套专为 AI 编程助手（如 Claude Code、Codex、Cursor 等）打造的高性能优化系统。它不仅仅是一组配置文件，而是一个经过长期实战打磨的完整框架，旨在解决 AI 代理在实际开发中面临的效率低下、记忆丢失、安全隐患及缺乏持续学习能力等核心痛点。\n\n通过引入技能模块化、直觉增强、记忆持久化机制以及内置的安全扫描功能，everything-claude-code 能显著提升 AI 在复杂任务中的表现，帮助开发者构建更稳定、更智能的生产级 AI 代理。其独特的“研究优先”开发理念和针对 Token 消耗的优化策略，使得模型响应更快、成本更低，同时有效防御潜在的攻击向量。\n\n这套工具特别适合软件开发者、AI 研究人员以及希望深度定制 AI 工作流的技术团队使用。无论您是在构建大型代码库，还是需要 AI 协助进行安全审计与自动化测试，everything-claude-code 都能提供强大的底层支持。作为一个曾荣获 Anthropic 黑客大奖的开源项目，它融合了多语言支持与丰富的实战钩子（hooks），让 AI 真正成长为懂上",138956,2,"2026-04-05T11:33:21",[13,15,26],"语言模型",{"id":28,"name":29,"github_repo":30,"description_zh":31,"stars":32,"difficulty_score":23,"last_commit_at":33,"category_tags":34,"status":16},2271,"ComfyUI","Comfy-Org\u002FComfyUI","ComfyUI 是一款功能强大且高度模块化的视觉 AI 引擎，专为设计和执行复杂的 Stable Diffusion 图像生成流程而打造。它摒弃了传统的代码编写模式，采用直观的节点式流程图界面，让用户通过连接不同的功能模块即可构建个性化的生成管线。\n\n这一设计巧妙解决了高级 AI 绘图工作流配置复杂、灵活性不足的痛点。用户无需具备编程背景，也能自由组合模型、调整参数并实时预览效果，轻松实现从基础文生图到多步骤高清修复等各类复杂任务。ComfyUI 拥有极佳的兼容性，不仅支持 Windows、macOS 和 Linux 全平台，还广泛适配 NVIDIA、AMD、Intel 及苹果 Silicon 等多种硬件架构，并率先支持 SDXL、Flux、SD3 等前沿模型。\n\n无论是希望深入探索算法潜力的研究人员和开发者，还是追求极致创作自由度的设计师与资深 AI 绘画爱好者，ComfyUI 都能提供强大的支持。其独特的模块化架构允许社区不断扩展新功能，使其成为当前最灵活、生态最丰富的开源扩散模型工具之一，帮助用户将创意高效转化为现实。",107662,"2026-04-03T11:11:01",[13,14,15],{"id":36,"name":37,"github_repo":38,"description_zh":39,"stars":40,"difficulty_score":23,"last_commit_at":41,"category_tags":42,"status":16},3704,"NextChat","ChatGPTNextWeb\u002FNextChat","NextChat 是一款轻量且极速的 AI 助手，旨在为用户提供流畅、跨平台的大模型交互体验。它完美解决了用户在多设备间切换时难以保持对话连续性，以及面对众多 AI 模型不知如何统一管理的痛点。无论是日常办公、学习辅助还是创意激发，NextChat 都能让用户随时随地通过网页、iOS、Android、Windows、MacOS 或 Linux 端无缝接入智能服务。\n\n这款工具非常适合普通用户、学生、职场人士以及需要私有化部署的企业团队使用。对于开发者而言，它也提供了便捷的自托管方案，支持一键部署到 Vercel 或 Zeabur 等平台。\n\nNextChat 的核心亮点在于其广泛的模型兼容性，原生支持 Claude、DeepSeek、GPT-4 及 Gemini Pro 等主流大模型，让用户在一个界面即可自由切换不同 AI 能力。此外，它还率先支持 MCP（Model Context Protocol）协议，增强了上下文处理能力。针对企业用户，NextChat 提供专业版解决方案，具备品牌定制、细粒度权限控制、内部知识库整合及安全审计等功能，满足公司对数据隐私和个性化管理的高标准要求。",87618,"2026-04-05T07:20:52",[13,26],{"id":44,"name":45,"github_repo":46,"description_zh":47,"stars":48,"difficulty_score":23,"last_commit_at":49,"category_tags":50,"status":16},2268,"ML-For-Beginners","microsoft\u002FML-For-Beginners","ML-For-Beginners 是由微软推出的一套系统化机器学习入门课程，旨在帮助零基础用户轻松掌握经典机器学习知识。这套课程将学习路径规划为 12 周，包含 26 节精炼课程和 52 道配套测验，内容涵盖从基础概念到实际应用的完整流程，有效解决了初学者面对庞大知识体系时无从下手、缺乏结构化指导的痛点。\n\n无论是希望转型的开发者、需要补充算法背景的研究人员，还是对人工智能充满好奇的普通爱好者，都能从中受益。课程不仅提供了清晰的理论讲解，还强调动手实践，让用户在循序渐进中建立扎实的技能基础。其独特的亮点在于强大的多语言支持，通过自动化机制提供了包括简体中文在内的 50 多种语言版本，极大地降低了全球不同背景用户的学习门槛。此外，项目采用开源协作模式，社区活跃且内容持续更新，确保学习者能获取前沿且准确的技术资讯。如果你正寻找一条清晰、友好且专业的机器学习入门之路，ML-For-Beginners 将是理想的起点。",84991,"2026-04-05T10:45:23",[14,51,52,53,15,54,26,13,55],"数据工具","视频","插件","其他","音频",{"id":57,"name":58,"github_repo":59,"description_zh":60,"stars":61,"difficulty_score":10,"last_commit_at":62,"category_tags":63,"status":16},3128,"ragflow","infiniflow\u002Fragflow","RAGFlow 是一款领先的开源检索增强生成（RAG）引擎，旨在为大语言模型构建更精准、可靠的上下文层。它巧妙地将前沿的 RAG 技术与智能体（Agent）能力相结合，不仅支持从各类文档中高效提取知识，还能让模型基于这些知识进行逻辑推理和任务执行。\n\n在大模型应用中，幻觉问题和知识滞后是常见痛点。RAGFlow 通过深度解析复杂文档结构（如表格、图表及混合排版），显著提升了信息检索的准确度，从而有效减少模型“胡编乱造”的现象，确保回答既有据可依又具备时效性。其内置的智能体机制更进一步，使系统不仅能回答问题，还能自主规划步骤解决复杂问题。\n\n这款工具特别适合开发者、企业技术团队以及 AI 研究人员使用。无论是希望快速搭建私有知识库问答系统，还是致力于探索大模型在垂直领域落地的创新者，都能从中受益。RAGFlow 提供了可视化的工作流编排界面和灵活的 API 接口，既降低了非算法背景用户的上手门槛，也满足了专业开发者对系统深度定制的需求。作为基于 Apache 2.0 协议开源的项目，它正成为连接通用大模型与行业专有知识之间的重要桥梁。",77062,"2026-04-04T04:44:48",[15,14,13,26,54],{"id":65,"github_repo":66,"name":67,"description_en":68,"description_zh":69,"ai_summary_zh":70,"readme_en":71,"readme_zh":72,"quickstart_zh":73,"use_case_zh":74,"hero_image_url":75,"owner_login":76,"owner_name":77,"owner_avatar_url":78,"owner_bio":79,"owner_company":80,"owner_location":80,"owner_email":80,"owner_twitter":80,"owner_website":81,"owner_url":82,"languages":83,"stars":98,"forks":99,"last_commit_at":100,"license":101,"difficulty_score":23,"env_os":102,"env_gpu":103,"env_ram":104,"env_deps":105,"category_tags":113,"github_topics":114,"view_count":23,"oss_zip_url":80,"oss_zip_packed_at":80,"status":16,"created_at":121,"updated_at":122,"faqs":123,"releases":149},2432,"google-deepmind\u002Fdm_control","dm_control","Google DeepMind's software stack for physics-based simulation and Reinforcement Learning environments, using MuJoCo.","dm_control 是 Google DeepMind 推出的一款基于 MuJoCo 物理引擎的开源软件栈，专为物理仿真和强化学习环境设计。它旨在为研究人员和开发者提供一套高效、标准化的工具，用于构建和测试涉及连续控制任务的智能体算法。\n\n在强化学习领域，构建高保真的物理模拟环境往往复杂且耗时。dm_control 通过封装底层的 MuJoCo 引擎，解决了这一痛点。它不仅提供了 Python 接口来直接调用物理引擎，还内置了一套丰富的标准强化学习环境套件（suite），涵盖了从简单的机械臂控制到复杂的多智能体足球对抗等多种场景。这使得用户无需从零开始搭建仿真世界，即可快速验证算法性能。\n\ndm_control 特别适合从事人工智能、机器人控制及强化学习研究的研究人员与工程师使用。对于希望深入理解物理交互机制或开发新型控制策略的开发者而言，它是一个不可或缺的基础设施。其独特的技术亮点在于模块化的设计：除了核心仿真功能，它还提供了 mjcf 库用于灵活修改模型结构，以及 composer 库允许用户像搭积木一样，通过复用独立组件来自定义复杂的交互环境。此外，dm_control 支持多","dm_control 是 Google DeepMind 推出的一款基于 MuJoCo 物理引擎的开源软件栈，专为物理仿真和强化学习环境设计。它旨在为研究人员和开发者提供一套高效、标准化的工具，用于构建和测试涉及连续控制任务的智能体算法。\n\n在强化学习领域，构建高保真的物理模拟环境往往复杂且耗时。dm_control 通过封装底层的 MuJoCo 引擎，解决了这一痛点。它不仅提供了 Python 接口来直接调用物理引擎，还内置了一套丰富的标准强化学习环境套件（suite），涵盖了从简单的机械臂控制到复杂的多智能体足球对抗等多种场景。这使得用户无需从零开始搭建仿真世界，即可快速验证算法性能。\n\ndm_control 特别适合从事人工智能、机器人控制及强化学习研究的研究人员与工程师使用。对于希望深入理解物理交互机制或开发新型控制策略的开发者而言，它是一个不可或缺的基础设施。其独特的技术亮点在于模块化的设计：除了核心仿真功能，它还提供了 mjcf 库用于灵活修改模型结构，以及 composer 库允许用户像搭积木一样，通过复用独立组件来自定义复杂的交互环境。此外，dm_control 支持多种渲染后端，既能满足本地调试时的可视化需求，也能在无头服务器上进行高性能的并行训练。总之，dm_control 以其严谨的工程实现和丰富的功能组件，成为了连接物理仿真与智能决策算法的重要桥梁。","# `dm_control`: Google DeepMind Infrastructure for Physics-Based Simulation.\n\nGoogle DeepMind's software stack for physics-based simulation and Reinforcement\nLearning environments, using MuJoCo physics.\n\nAn **introductory tutorial** for this package is available as a Colaboratory\nnotebook:\n[![Open In Colab](https:\u002F\u002Fcolab.research.google.com\u002Fassets\u002Fcolab-badge.svg)](https:\u002F\u002Fcolab.research.google.com\u002Fgithub\u002Fgoogle-deepmind\u002Fdm_control\u002Fblob\u002Fmain\u002Ftutorial.ipynb)\n\n## Overview\n\nThis package consists of the following \"core\" components:\n\n-   [`dm_control.mujoco`]: Libraries that provide Python bindings to the MuJoCo\n    physics engine.\n\n-   [`dm_control.suite`]: A set of Python Reinforcement Learning environments\n    powered by the MuJoCo physics engine.\n\n-   [`dm_control.viewer`]: An interactive environment viewer.\n\nAdditionally, the following components are available for the creation of more\ncomplex control tasks:\n\n-   [`dm_control.mjcf`]: A library for composing and modifying MuJoCo MJCF\n    models in Python.\n\n-   `dm_control.composer`: A library for defining rich RL environments from\n    reusable, self-contained components.\n\n-   [`dm_control.locomotion`]: Additional libraries for custom tasks.\n\n-   [`dm_control.locomotion.soccer`]: Multi-agent soccer tasks.\n\nIf you use this package, please cite our accompanying [publication]:\n\n```\n@article{tunyasuvunakool2020,\n         title = {dm_control: Software and tasks for continuous control},\n         journal = {Software Impacts},\n         volume = {6},\n         pages = {100022},\n         year = {2020},\n         issn = {2665-9638},\n         doi = {https:\u002F\u002Fdoi.org\u002F10.1016\u002Fj.simpa.2020.100022},\n         url = {https:\u002F\u002Fwww.sciencedirect.com\u002Fscience\u002Farticle\u002Fpii\u002FS2665963820300099},\n         author = {Saran Tunyasuvunakool and Alistair Muldal and Yotam Doron and\n                   Siqi Liu and Steven Bohez and Josh Merel and Tom Erez and\n                   Timothy Lillicrap and Nicolas Heess and Yuval Tassa},\n}\n```\n\n## Installation\n\nInstall `dm_control` from PyPI by running\n\n```sh\npip install dm_control\n```\n\n> **Note**: **`dm_control` cannot be installed in \"editable\" mode** (i.e. `pip\n> install -e`).\n>\n> While `dm_control` has been largely updated to use the pybind11-based bindings\n> provided via the `mujoco` package, at this time it still relies on some legacy\n> components that are automatically generated from MuJoCo header files in a way\n> that is incompatible with editable mode. Attempting to install `dm_control` in\n> editable mode will result in import errors like:\n>\n> ```\n> ImportError: cannot import name 'constants' from partially initialized module 'dm_control.mujoco.wrapper.mjbindings' ...\n> ```\n>\n> The solution is to `pip uninstall dm_control` and then reinstall it without\n> the `-e` flag.\n\n## Versioning\n\nStarting from version 1.0.0, we adopt semantic versioning.\n\nPrior to version 1.0.0, the `dm_control` Python package was versioned `0.0.N`,\nwhere `N` was an internal revision number that increased by an arbitrary amount\nat every single Git commit.\n\nIf you want to install an unreleased version of `dm_control` directly from our\nrepository, you can do so by running `pip install\ngit+https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control.git`.\n\n## Rendering\n\nThe MuJoCo Python bindings support three different OpenGL rendering backends:\nEGL (headless, hardware-accelerated), GLFW (windowed, hardware-accelerated), and\nOSMesa (purely software-based). At least one of these three backends must be\navailable in order render through `dm_control`.\n\n*   Hardware rendering with a windowing system is supported via GLFW and GLEW.\n    On Linux these can be installed using your distribution's package manager.\n    For example, on Debian and Ubuntu, this can be done by running `sudo apt-get\n    install libglfw3 libglew2.0`. Please note that:\n\n    -   [`dm_control.viewer`] can only be used with GLFW.\n    -   GLFW will not work on headless machines.\n\n*   \"Headless\" hardware rendering (i.e. without a windowing system such as X11)\n    requires [EXT_platform_device] support in the EGL driver. Recent Nvidia\n    drivers support this. You will also need GLEW. On Debian and Ubuntu, this\n    can be installed via `sudo apt-get install libglew2.0`.\n\n*   Software rendering requires GLX and OSMesa. On Debian and Ubuntu these can\n    be installed using `sudo apt-get install libgl1-mesa-glx libosmesa6`.\n\nBy default, `dm_control` will attempt to use GLFW first, then EGL, then OSMesa.\nYou can also specify a particular backend to use by setting the `MUJOCO_GL=`\nenvironment variable to `\"glfw\"`, `\"egl\"`, or `\"osmesa\"`, respectively. When\nrendering with EGL, you can also specify which GPU to use for rendering by\nsetting the environment variable `MUJOCO_EGL_DEVICE_ID=` to the target GPU ID.\n\n## Additional instructions for Homebrew users on macOS\n\n1.  The above instructions using `pip` should work, provided that you use a\n    Python interpreter that is installed by Homebrew (rather than the\n    system-default one).\n\n2.  Before running, the `DYLD_LIBRARY_PATH` environment variable needs to be\n    updated with the path to the GLFW library. This can be done by running\n    `export DYLD_LIBRARY_PATH=$(brew --prefix)\u002Flib:$DYLD_LIBRARY_PATH`.\n\n[EXT_platform_device]: https:\u002F\u002Fwww.khronos.org\u002Fregistry\u002FEGL\u002Fextensions\u002FEXT\u002FEGL_EXT_platform_device.txt\n[Releases page on the MuJoCo GitHub repository]: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fmujoco\u002Freleases\n[MuJoCo website]: https:\u002F\u002Fmujoco.org\u002F\n[publication]: https:\u002F\u002Fdoi.org\u002F10.1016\u002Fj.simpa.2020.100022\n[`ctypes`]: https:\u002F\u002Fdocs.python.org\u002F3\u002Flibrary\u002Fctypes.html\n[`dm_control.mjcf`]: dm_control\u002Fmjcf\u002FREADME.md\n[`dm_control.mujoco`]: dm_control\u002Fmujoco\u002FREADME.md\n[`dm_control.suite`]: dm_control\u002Fsuite\u002FREADME.md\n[`dm_control.viewer`]: dm_control\u002Fviewer\u002FREADME.md\n[`dm_control.locomotion`]: dm_control\u002Flocomotion\u002FREADME.md\n[`dm_control.locomotion.soccer`]: dm_control\u002Flocomotion\u002Fsoccer\u002FREADME.md\n","# `dm_control`: 基于物理模拟的 Google DeepMind 基础设施。\n\nGoogle DeepMind 用于基于物理的仿真和强化学习环境的软件栈，采用 MuJoCo 物理引擎。\n\n该包的**入门教程**以 Colaboratory 笔记本形式提供：\n[![在 Colab 中打开](https:\u002F\u002Fcolab.research.google.com\u002Fassets\u002Fcolab-badge.svg)](https:\u002F\u002Fcolab.research.google.com\u002Fgithub\u002Fgoogle-deepmind\u002Fdm_control\u002Fblob\u002Fmain\u002Ftutorial.ipynb)\n\n## 概述\n\n本包由以下“核心”组件组成：\n\n-   [`dm_control.mujoco`]: 提供 Python 绑定到 MuJoCo 物理引擎的库。\n\n-   [`dm_control.suite`]: 一组由 MuJoCo 物理引擎驱动的 Python 强化学习环境。\n\n-   [`dm_control.viewer`]: 一个交互式环境查看器。\n\n此外，还提供了以下组件，用于创建更复杂的控制任务：\n\n-   [`dm_control.mjcf`]: 一个用于在 Python 中组合和修改 MuJoCo MJCF 模型的库。\n\n-   `dm_control.composer`: 一个用于从可重用、自包含的组件中定义丰富 RL 环境的库。\n\n-   [`dm_control.locomotion`]: 用于自定义任务的附加库。\n\n-   [`dm_control.locomotion.soccer`]: 多智能体足球任务。\n\n如果您使用本包，请引用我们的配套[论文]：\n\n```\n@article{tunyasuvunakool2020,\n         title = {dm_control: Software and tasks for continuous control},\n         journal = {Software Impacts},\n         volume = {6},\n         pages = {100022},\n         year = {2020},\n         issn = {2665-9638},\n         doi = {https:\u002F\u002Fdoi.org\u002F10.1016\u002Fj.simpa.2020.100022},\n         url = {https:\u002F\u002Fwww.sciencedirect.com\u002Fscience\u002Farticle\u002Fpii\u002FS2665963820300099},\n         author = {Saran Tunyasuvunakool and Alistair Muldal and Yotam Doron and\n                   Siqi Liu and Steven Bohez and Josh Merel and Tom Erez and\n                   Timothy Lillicrap and Nicolas Heess and Yuval Tassa},\n}\n```\n\n## 安装\n\n通过运行以下命令从 PyPI 安装 `dm_control`：\n\n```sh\npip install dm_control\n```\n\n> **注意**：**`dm_control` 不能以“可编辑”模式安装**（即 `pip install -e`）。\n>\n> 虽然 `dm_control` 已经大部分更新为使用通过 `mujoco` 包提供的基于 pybind11 的绑定，但目前它仍然依赖于一些从 MuJoCo 头文件自动生成的旧版组件，这些组件与可编辑模式不兼容。尝试以可编辑模式安装 `dm_control` 将导致类似以下的导入错误：\n>\n> ```\n> ImportError: cannot import name 'constants' from partially initialized module 'dm_control.mujoco.wrapper.mjbindings' ...\n> ```\n>\n> 解决方法是先卸载 `dm_control`，然后不带 `-e` 标志重新安装。\n\n## 版本管理\n\n自 1.0.0 版本起，我们采用语义版本控制。\n\n在 1.0.0 版本之前，`dm_control` Python 包的版本号为 `0.0.N`，其中 `N` 是一个内部修订号，在每次 Git 提交时都会任意递增。\n\n如果您想直接从我们的仓库安装尚未发布的 `dm_control` 版本，可以运行 `pip install git+https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control.git`。\n\n## 渲染\n\nMuJoCo 的 Python 绑定支持三种不同的 OpenGL 渲染后端：EGL（无头、硬件加速）、GLFW（窗口化、硬件加速）和 OSMesa（纯软件）。必须至少有一个后端可用，才能通过 `dm_control` 进行渲染。\n\n*   使用窗口系统的硬件渲染由 GLFW 和 GLEW 支持。在 Linux 上，可以通过发行版的包管理器安装它们。例如，在 Debian 和 Ubuntu 上，可以运行 `sudo apt-get install libglfw3 libglew2.0`。请注意：\n\n    -   [`dm_control.viewer`] 只能与 GLFW 一起使用。\n    -   GLFW 在无头机器上无法工作。\n\n*   “无头”硬件渲染（即没有 X11 等窗口系统）需要 EGL 驱动程序支持 [EXT_platform_device] 扩展。最近的 Nvidia 驱动程序支持此功能。您还需要 GLEW。在 Debian 和 Ubuntu 上，可以通过 `sudo apt-get install libglew2.0` 安装。\n\n*   软件渲染需要 GLX 和 OSMesa。在 Debian 和 Ubuntu 上，可以通过 `sudo apt-get install libgl1-mesa-glx libosmesa6` 安装。\n\n默认情况下，`dm_control` 会优先尝试使用 GLFW，然后是 EGL，最后是 OSMesa。您也可以通过将 `MUJOCO_GL=` 环境变量分别设置为 `\"glfw\"`、`\"egl\"` 或 `\"osmesa\"` 来指定要使用的特定后端。使用 EGL 渲染时，还可以通过设置环境变量 `MUJOCO_EGL_DEVICE_ID=` 来指定要使用的 GPU ID。\n\n## macOS 上 Homebrew 用户的补充说明\n\n1.  上述使用 `pip` 的说明应有效，前提是您使用的是由 Homebrew 安装的 Python 解释器（而非系统默认的解释器）。\n\n2.  在运行之前，需要将 `DYLD_LIBRARY_PATH` 环境变量更新为 GLFW 库的路径。这可以通过运行 `export DYLD_LIBRARY_PATH=$(brew --prefix)\u002Flib:$DYLD_LIBRARY_PATH` 来完成。\n\n[EXT_platform_device]: https:\u002F\u002Fwww.khronos.org\u002Fregistry\u002FEGL\u002Fextensions\u002FEXT\u002FEGL_EXT_platform_device.txt\n[MuJoCo GitHub 仓库的发布页面]: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fmujoco\u002Freleases\n[MuJoCo 官网]: https:\u002F\u002Fmujoco.org\u002F\n[论文]: https:\u002F\u002Fdoi.org\u002F10.1016\u002Fj.simpa.2020.100022\n[`ctypes`]: https:\u002F\u002Fdocs.python.org\u002F3\u002Flibrary\u002Fctypes.html\n[`dm_control.mjcf`]: dm_control\u002Fmjcf\u002FREADME.md\n[`dm_control.mujoco`]: dm_control\u002Fmujoco\u002FREADME.md\n[`dm_control.suite`]: dm_control\u002Fsuite\u002FREADME.md\n[`dm_control.viewer`]: dm_control\u002Fviewer\u002FREADME.md\n[`dm_control.locomotion`]: dm_control\u002Flocomotion\u002FREADME.md\n[`dm_control.locomotion.soccer`]: dm_control\u002Flocomotion\u002Fsoccer\u002FREADME.md","# dm_control 快速上手指南\n\n`dm_control` 是 Google DeepMind 提供的基于 MuJoCo 物理引擎的仿真软件栈，主要用于强化学习环境构建和连续控制任务。\n\n## 环境准备\n\n### 系统要求与依赖\n`dm_control` 支持 Linux、macOS 和 Windows。渲染功能依赖 OpenGL 后端，至少需要安装以下三者之一：\n1. **GLFW**（窗口化，硬件加速）：适用于本地开发调试。\n2. **EGL**（无头模式，硬件加速）：适用于服务器或无显示环境。\n3. **OSMesa**（纯软件渲染）：适用于无 GPU 环境。\n\n### Linux (Debian\u002FUbuntu) 前置依赖安装\n根据选择的渲染后端，运行以下命令安装系统库：\n\n*   **使用 GLFW (推荐本地开发):**\n    ```sh\n    sudo apt-get install libglfw3 libglew2.0\n    ```\n*   **使用 EGL (无头硬件加速):**\n    ```sh\n    sudo apt-get install libglew2.0\n    ```\n    *注意：需要 NVIDIA 驱动支持 `EXT_platform_device`。*\n*   **使用 OSMesa (软件渲染):**\n    ```sh\n    sudo apt-get install libgl1-mesa-glx libosmesa6\n    ```\n\n### macOS 用户额外配置\n如果使用 Homebrew 安装的 Python，需设置环境变量以指向 GLFW 库路径：\n```sh\nexport DYLD_LIBRARY_PATH=$(brew --prefix)\u002Flib:$DYLD_LIBRARY_PATH\n```\n\n## 安装步骤\n\n通过 PyPI 直接安装 `dm_control`。\n\n```sh\npip install dm_control\n```\n\n> **重要提示**：**不支持**以“可编辑模式”安装（即不要使用 `pip install -e .`）。由于部分遗留组件的生成方式限制，使用 `-e` 标志会导致导入错误。如需从源码安装最新未发布版本，请使用：\n> ```sh\n> pip install git+https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control.git\n> ```\n\n## 基本使用\n\n### 1. 快速体验教程\n官方提供了 Colab 笔记本教程，适合初学者快速了解核心功能：\n[在 Colab 中打开教程](https:\u002F\u002Fcolab.research.google.com\u002Fgithub\u002Fgoogle-deepmind\u002Fdm_control\u002Fblob\u002Fmain\u002Ftutorial.ipynb)\n\n### 2. 代码示例\n以下是一个加载标准强化学习环境并运行一步仿真的最小示例：\n\n```python\nfrom dm_control import suite\nimport numpy as np\n\n# 加载一个标准的强化学习环境 (例如: Cartpole 平衡任务)\nenv = suite.load(domain_name=\"cartpole\", task_name=\"balance\")\n\n# 获取初始观测\ntimestep = env.reset()\n\n# 执行一个随机动作\naction = np.random.uniform(-1, 1, env.action_spec().shape)\ntimestep = env.step(action)\n\n# 输出观测信息\nprint(\"Observation:\", timestep.observation)\nprint(\"Reward:\", timestep.reward)\nprint(\"Discount:\", timestep.discount)\n```\n\n### 3. 使用交互式查看器\n`dm_control` 内置了一个交互式查看器，可用于可视化仿真过程（需 GLFW 支持）：\n\n```python\nfrom dm_control import viewer\nfrom dm_control import suite\n\n# 定义一个简单的策略函数（此处为随机策略）\ndef policy(timestep):\n    return np.random.uniform(-1, 1, size=env.action_spec().shape)\n\n# 加载环境\nenv = suite.load(domain_name=\"humanoid\", task_name=\"stand\")\n\n# 启动查看器\nviewer.launch(env, policy=policy)\n```\n\n### 4. 指定渲染后端\n默认情况下，`dm_control` 会依次尝试 GLFW -> EGL -> OSMesa。你可以通过设置环境变量强制指定后端：\n\n```sh\n# 强制使用 EGL (适用于无头服务器)\nexport MUJOCO_GL=egl\n\n# 强制使用 GLFW\nexport MUJOCO_GL=glfw\n\n# 强制使用 OSMesa\nexport MUJOCO_GL=osmesa\n```\n\n若使用 EGL 且有多块 GPU，可通过 `MUJOCO_EGL_DEVICE_ID` 指定使用的 GPU ID。","某机器人算法团队正在研发一款双足行走机器人的平衡控制策略，需要构建高保真的物理仿真环境来训练强化学习代理，以替代昂贵且高风险的真机试错过程。\n\n### 没有 dm_control 时\n- **物理引擎集成繁琐**：开发者需手动编写 Python 与 MuJoCo C++ 底层接口的绑定代码，配置复杂的编译环境，极易出现版本冲突或导入错误，初期搭建耗时数周。\n- **场景构建效率低下**：修改机器人模型或环境参数需直接编辑晦涩的 XML 文件，缺乏程序化生成能力，难以快速迭代不同的地形或负载测试用例。\n- **调试可视化困难**：缺乏原生的交互式查看器，开发者只能依赖静态截图或自行集成第三方渲染后端，无法实时观察关节力矩和接触力变化，排查“仿真崩溃”问题如同盲人摸象。\n- **基准对比缺失**：需从零实现基础控制任务（如站立、行走），难以验证算法的有效性，且无法与业界标准进行公平的性能对标。\n\n### 使用 dm_control 后\n- **开箱即用的集成体验**：通过 `pip install` 即可直接调用经过优化的 MuJoCo Python 绑定，无需处理底层编译细节，将环境搭建时间从数周缩短至几分钟。\n- **程序化建模与组合**：利用 `dm_control.mjcf` 和 `composer` 库，开发者可用 Python 代码动态组装机器人组件和环境元素，轻松实现参数化扫描和复杂场景的自动化生成。\n- **原生交互与调试支持**：内置的 `dm_control.viewer` 提供流畅的交互式可视化界面，支持实时调整相机视角和物理参数，结合 EGL\u002FGLFW 后端，让开发者能直观监控仿真状态并快速定位异常。\n- **丰富的标准任务套件**：直接调用 `dm_control.suite` 中预置的标准化控制任务（如 walker、humanoid），不仅提供了可靠的基线性能参考，还确保了实验结果的可复现性和学术可比性。\n\n核心价值在于，dm_control 将复杂的物理仿真工程难题封装为简洁的 Python API，让研究人员能从底层基建中解放出来，专注于核心控制算法的创新与迭代。","https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002Fgoogle-deepmind_dm_control_c601d84f.png","google-deepmind","Google DeepMind","https:\u002F\u002Foss.gittoolsai.com\u002Favatars\u002Fgoogle-deepmind_06b1dd17.png","",null,"https:\u002F\u002Fwww.deepmind.com\u002F","https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind",[84,88,92,96],{"name":85,"color":86,"percentage":87},"Python","#3572A5",94.4,{"name":89,"color":90,"percentage":91},"Jupyter Notebook","#DA5B0B",5.6,{"name":93,"color":94,"percentage":95},"Shell","#89e051",0,{"name":97,"color":80,"percentage":95},"Linker Script",4522,744,"2026-04-02T08:36:28","Apache-2.0","Linux, macOS","非必需。若需硬件加速渲染，支持 NVIDIA GPU（需 EGL 驱动支持 EXT_platform_device）或任意支持 OpenGL 的显卡；也可使用纯软件渲染（OSMesa）。","未说明",{"notes":106,"python":104,"dependencies":107},"1. 不支持以“可编辑模式”（pip install -e）安装，否则会导致导入错误。\n2. 渲染后端至少需要以下三者之一：GLFW（窗口化硬件加速）、EGL（无头硬件加速）或 OSMesa（纯软件渲染）。\n3. Linux 下需安装系统级依赖：libglfw3, libglew2.0, libgl1-mesa-glx, libosmesa6。\n4. macOS Homebrew 用户需使用 Homebrew 安装的 Python，并设置 DYLD_LIBRARY_PATH 环境变量指向 GLFW 库路径。\n5. 可通过环境变量 MUJOCO_GL 指定渲染后端（glfw\u002Fegl\u002Fosmesa），使用 EGL 时可通过 MUJOCO_EGL_DEVICE_ID 指定 GPU。",[108,109,110,111,112],"mujoco","pybind11","glfw","glew","libosmesa6",[13,54],[115,116,117,118,119,120,108],"machine-learning","artificial-intelligence","neural-networks","deep-learning","reinforcement-learning","physics-simulation","2026-03-27T02:49:30.150509","2026-04-06T05:19:27.115859",[124,129,134,139,144],{"id":125,"question_zh":126,"answer_zh":127,"source_url":128},11191,"在 macOS 上使用 Conda 安装 dm_control 时遇到 'Unable to load EGL library' 或 glfw 导入错误怎么办？","如果在 macOS 上通过 Conda 环境运行 dm_control 并遇到 EGL 库加载失败或 glfw 相关的 MemoryError，建议从 conda-forge 渠道安装 pyglfw。执行命令：conda install -c conda-forge pyglfw。这通常能解决依赖冲突问题。","https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fissues\u002F223",{"id":130,"question_zh":131,"answer_zh":132,"source_url":133},11192,"导入 dm_control 时出现 \"ImportError: cannot import name 'constants'\" 错误如何解决？","这通常是因为你的项目目录名称与库名称 \"dm_control\" 冲突（例如当前文件夹名为 dm_control），导致 Python 导入了本地目录而非安装的库。解决方法是将你的项目目录重命名为其他名称（例如 dm_control1 或 my_project），避免命名冲突。","https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fissues\u002F6",{"id":135,"question_zh":136,"answer_zh":137,"source_url":138},11193,"使用 MuJoCo 渲染时图像出现噪点、花屏或不可读的情况怎么办？","这是一个已知的渲染 Bug，特别是在使用 EGL 后端进行并行环境渲染时。该问题已在 dm_control 版本 1.0.6 中修复。请升级你的 dm_control 库到 1.0.6 或更高版本以解决此问题。","https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fissues\u002F310",{"id":140,"question_zh":141,"answer_zh":142,"source_url":143},11194,"遇到 \"mujoco.FatalError: gladLoadGL error\" 错误该如何修复？","这个错误通常与 OpenGL 上下文加载有关。可以参考 MuJoCo 官方仓库的相关讨论（google-deepmind\u002Fmujoco#572）。常见的解决方法包括检查显卡驱动、确保正确设置了 MUJOCO_GL 环境变量，或者在无头服务器（Headless Server）上尝试不同的渲染后端配置。","https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fissues\u002F283",{"id":145,"question_zh":146,"answer_zh":147,"source_url":148},11195,"在无头服务器（Headless Server）或集群上使用 EGL 渲染时遇到初始化错误或状态异常怎么办？","EGL 后端的错误有时与机器状态有关。如果遇到奇怪的初始化错误，尝试重启机器可能暂时解决问题。此外，确保你有足够的权限（sudo 访问权可能有助于 Conda 包的正确链接），并参考 PyTorch RL 等文档中关于 MuJoCo 安装的常见问题的建议。如果通过 SSH 运行，请确保正确配置了无头显示后端。","https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fissues\u002F123",[150,155,160,165,170,175,180,185,190,195,200,205,210,215,220,225,230,235,239,243],{"id":151,"version":152,"summary_zh":153,"released_at":154},61650,"1.0.38","**完整更新日志**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.37...1.0.38","2026-03-11T08:34:39",{"id":156,"version":157,"summary_zh":158,"released_at":159},61651,"1.0.37","**完整更新日志**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.36...1.0.37","2026-02-13T10:54:46",{"id":161,"version":162,"summary_zh":163,"released_at":164},61652,"1.0.36","**完整更新日志**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.34...1.0.36","2026-01-08T12:37:27",{"id":166,"version":167,"summary_zh":168,"released_at":169},61653,"1.0.34","**完整更新日志**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.31...1.0.34","2025-09-19T21:15:47",{"id":171,"version":172,"summary_zh":173,"released_at":174},61654,"1.0.31","**完整变更日志**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.30...1.0.31","2025-06-11T21:11:49",{"id":176,"version":177,"summary_zh":178,"released_at":179},61655,"1.0.30","**完整变更日志**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.28...1.0.30","2025-04-30T16:15:49",{"id":181,"version":182,"summary_zh":183,"released_at":184},61656,"1.0.28","**完整更新日志**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.26...1.0.28","2025-03-06T17:46:01",{"id":186,"version":187,"summary_zh":188,"released_at":189},61657,"1.0.26","**完整变更日志**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.25...1.0.26","2024-12-06T18:03:46",{"id":191,"version":192,"summary_zh":193,"released_at":194},61658,"1.0.25","**完整更新日志**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.24...1.0.25","2024-11-05T18:39:22",{"id":196,"version":197,"summary_zh":198,"released_at":199},61659,"1.0.24","**完整变更日志**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.23...1.0.24","2024-10-23T13:34:43",{"id":201,"version":202,"summary_zh":203,"released_at":204},61660,"1.0.23","**Full Changelog**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.22...1.0.23","2024-10-04T09:55:38",{"id":206,"version":207,"summary_zh":208,"released_at":209},61661,"1.0.22","Bump dm_control version to 1.0.22.\r\n\r\nDepend on mujoco 3.2.1.\r\n\r\n**Full Changelog**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.21...1.0.22","2024-08-07T15:44:06",{"id":211,"version":212,"summary_zh":213,"released_at":214},61662,"1.0.21","Bump dm_control version to 1.0.21.\r\n\r\nDepend on mujoco 3.2.0.\r\n\r\n**Full Changelog**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.20...1.0.21","2024-07-29T17:32:24",{"id":216,"version":217,"summary_zh":218,"released_at":219},61663,"1.0.20","Bump dm_control version to 1.0.20.\r\n\r\nDepend on mujoco 3.1.6.\r\n\r\n**Full Changelog**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.19...1.0.20","2024-06-04T17:45:31",{"id":221,"version":222,"summary_zh":223,"released_at":224},61664,"1.0.19","Bump dm_control version to 1.0.19.\r\n\r\nDepend on mujoco 3.1.5.\r\n\r\n**Full Changelog**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.18...1.0.19","2024-05-08T09:41:25",{"id":226,"version":227,"summary_zh":228,"released_at":229},61665,"1.0.18","Bump dm_control version to 1.0.18.\r\n\r\nDepend on mujoco 3.1.4.\r\n\r\n**Full Changelog**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.17...1.0.18","2024-04-12T13:20:41",{"id":231,"version":232,"summary_zh":233,"released_at":234},61666,"1.0.17","Bump dm_control version to 1.0.17.\r\n\r\nDepend on mujoco 3.1.3.\r\n\r\n**Full Changelog**: https:\u002F\u002Fgithub.com\u002Fgoogle-deepmind\u002Fdm_control\u002Fcompare\u002F1.0.16...1.0.17","2024-03-29T09:17:04",{"id":236,"version":237,"summary_zh":80,"released_at":238},61667,"1.0.16","2023-12-20T12:21:44",{"id":240,"version":241,"summary_zh":80,"released_at":242},61668,"1.0.15","2023-10-18T16:13:08",{"id":244,"version":245,"summary_zh":80,"released_at":246},61669,"1.0.14","2023-07-20T12:03:39"]