[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"similar-kaito-project--aikit":3,"tool-kaito-project--aikit":61},[4,18,26,36,44,53],{"id":5,"name":6,"github_repo":7,"description_zh":8,"stars":9,"difficulty_score":10,"last_commit_at":11,"category_tags":12,"status":17},4358,"openclaw","openclaw\u002Fopenclaw","OpenClaw 是一款专为个人打造的本地化 AI 助手，旨在让你在自己的设备上拥有完全可控的智能伙伴。它打破了传统 AI 助手局限于特定网页或应用的束缚，能够直接接入你日常使用的各类通讯渠道，包括微信、WhatsApp、Telegram、Discord、iMessage 等数十种平台。无论你在哪个聊天软件中发送消息，OpenClaw 都能即时响应，甚至支持在 macOS、iOS 和 Android 设备上进行语音交互，并提供实时的画布渲染功能供你操控。\n\n这款工具主要解决了用户对数据隐私、响应速度以及“始终在线”体验的需求。通过将 AI 部署在本地，用户无需依赖云端服务即可享受快速、私密的智能辅助，真正实现了“你的数据，你做主”。其独特的技术亮点在于强大的网关架构，将控制平面与核心助手分离，确保跨平台通信的流畅性与扩展性。\n\nOpenClaw 非常适合希望构建个性化工作流的技术爱好者、开发者，以及注重隐私保护且不愿被单一生态绑定的普通用户。只要具备基础的终端操作能力（支持 macOS、Linux 及 Windows WSL2），即可通过简单的命令行引导完成部署。如果你渴望拥有一个懂你",349277,3,"2026-04-06T06:32:30",[13,14,15,16],"Agent","开发框架","图像","数据工具","ready",{"id":19,"name":20,"github_repo":21,"description_zh":22,"stars":23,"difficulty_score":10,"last_commit_at":24,"category_tags":25,"status":17},3808,"stable-diffusion-webui","AUTOMATIC1111\u002Fstable-diffusion-webui","stable-diffusion-webui 是一个基于 Gradio 构建的网页版操作界面，旨在让用户能够轻松地在本地运行和使用强大的 Stable Diffusion 图像生成模型。它解决了原始模型依赖命令行、操作门槛高且功能分散的痛点，将复杂的 AI 绘图流程整合进一个直观易用的图形化平台。\n\n无论是希望快速上手的普通创作者、需要精细控制画面细节的设计师，还是想要深入探索模型潜力的开发者与研究人员，都能从中获益。其核心亮点在于极高的功能丰富度：不仅支持文生图、图生图、局部重绘（Inpainting）和外绘（Outpainting）等基础模式，还独创了注意力机制调整、提示词矩阵、负向提示词以及“高清修复”等高级功能。此外，它内置了 GFPGAN 和 CodeFormer 等人脸修复工具，支持多种神经网络放大算法，并允许用户通过插件系统无限扩展能力。即使是显存有限的设备，stable-diffusion-webui 也提供了相应的优化选项，让高质量的 AI 艺术创作变得触手可及。",162132,"2026-04-05T11:01:52",[14,15,13],{"id":27,"name":28,"github_repo":29,"description_zh":30,"stars":31,"difficulty_score":32,"last_commit_at":33,"category_tags":34,"status":17},1381,"everything-claude-code","affaan-m\u002Feverything-claude-code","everything-claude-code 是一套专为 AI 编程助手（如 Claude Code、Codex、Cursor 等）打造的高性能优化系统。它不仅仅是一组配置文件，而是一个经过长期实战打磨的完整框架，旨在解决 AI 代理在实际开发中面临的效率低下、记忆丢失、安全隐患及缺乏持续学习能力等核心痛点。\n\n通过引入技能模块化、直觉增强、记忆持久化机制以及内置的安全扫描功能，everything-claude-code 能显著提升 AI 在复杂任务中的表现，帮助开发者构建更稳定、更智能的生产级 AI 代理。其独特的“研究优先”开发理念和针对 Token 消耗的优化策略，使得模型响应更快、成本更低，同时有效防御潜在的攻击向量。\n\n这套工具特别适合软件开发者、AI 研究人员以及希望深度定制 AI 工作流的技术团队使用。无论您是在构建大型代码库，还是需要 AI 协助进行安全审计与自动化测试，everything-claude-code 都能提供强大的底层支持。作为一个曾荣获 Anthropic 黑客大奖的开源项目，它融合了多语言支持与丰富的实战钩子（hooks），让 AI 真正成长为懂上",156804,2,"2026-04-15T11:34:33",[14,13,35],"语言模型",{"id":37,"name":38,"github_repo":39,"description_zh":40,"stars":41,"difficulty_score":32,"last_commit_at":42,"category_tags":43,"status":17},2271,"ComfyUI","Comfy-Org\u002FComfyUI","ComfyUI 是一款功能强大且高度模块化的视觉 AI 引擎，专为设计和执行复杂的 Stable Diffusion 图像生成流程而打造。它摒弃了传统的代码编写模式，采用直观的节点式流程图界面，让用户通过连接不同的功能模块即可构建个性化的生成管线。\n\n这一设计巧妙解决了高级 AI 绘图工作流配置复杂、灵活性不足的痛点。用户无需具备编程背景，也能自由组合模型、调整参数并实时预览效果，轻松实现从基础文生图到多步骤高清修复等各类复杂任务。ComfyUI 拥有极佳的兼容性，不仅支持 Windows、macOS 和 Linux 全平台，还广泛适配 NVIDIA、AMD、Intel 及苹果 Silicon 等多种硬件架构，并率先支持 SDXL、Flux、SD3 等前沿模型。\n\n无论是希望深入探索算法潜力的研究人员和开发者，还是追求极致创作自由度的设计师与资深 AI 绘画爱好者，ComfyUI 都能提供强大的支持。其独特的模块化架构允许社区不断扩展新功能，使其成为当前最灵活、生态最丰富的开源扩散模型工具之一，帮助用户将创意高效转化为现实。",108322,"2026-04-10T11:39:34",[14,15,13],{"id":45,"name":46,"github_repo":47,"description_zh":48,"stars":49,"difficulty_score":32,"last_commit_at":50,"category_tags":51,"status":17},6121,"gemini-cli","google-gemini\u002Fgemini-cli","gemini-cli 是一款由谷歌推出的开源 AI 命令行工具，它将强大的 Gemini 大模型能力直接集成到用户的终端环境中。对于习惯在命令行工作的开发者而言，它提供了一条从输入提示词到获取模型响应的最短路径，无需切换窗口即可享受智能辅助。\n\n这款工具主要解决了开发过程中频繁上下文切换的痛点，让用户能在熟悉的终端界面内直接完成代码理解、生成、调试以及自动化运维任务。无论是查询大型代码库、根据草图生成应用，还是执行复杂的 Git 操作，gemini-cli 都能通过自然语言指令高效处理。\n\n它特别适合广大软件工程师、DevOps 人员及技术研究人员使用。其核心亮点包括支持高达 100 万 token 的超长上下文窗口，具备出色的逻辑推理能力；内置 Google 搜索、文件操作及 Shell 命令执行等实用工具；更独特的是，它支持 MCP（模型上下文协议），允许用户灵活扩展自定义集成，连接如图像生成等外部能力。此外，个人谷歌账号即可享受免费的额度支持，且项目基于 Apache 2.0 协议完全开源，是提升终端工作效率的理想助手。",100752,"2026-04-10T01:20:03",[52,13,15,14],"插件",{"id":54,"name":55,"github_repo":56,"description_zh":57,"stars":58,"difficulty_score":32,"last_commit_at":59,"category_tags":60,"status":17},4721,"markitdown","microsoft\u002Fmarkitdown","MarkItDown 是一款由微软 AutoGen 团队打造的轻量级 Python 工具，专为将各类文件高效转换为 Markdown 格式而设计。它支持 PDF、Word、Excel、PPT、图片（含 OCR）、音频（含语音转录）、HTML 乃至 YouTube 链接等多种格式的解析，能够精准提取文档中的标题、列表、表格和链接等关键结构信息。\n\n在人工智能应用日益普及的今天，大语言模型（LLM）虽擅长处理文本，却难以直接读取复杂的二进制办公文档。MarkItDown 恰好解决了这一痛点，它将非结构化或半结构化的文件转化为模型“原生理解”且 Token 效率极高的 Markdown 格式，成为连接本地文件与 AI 分析 pipeline 的理想桥梁。此外，它还提供了 MCP（模型上下文协议）服务器，可无缝集成到 Claude Desktop 等 LLM 应用中。\n\n这款工具特别适合开发者、数据科学家及 AI 研究人员使用，尤其是那些需要构建文档检索增强生成（RAG）系统、进行批量文本分析或希望让 AI 助手直接“阅读”本地文件的用户。虽然生成的内容也具备一定可读性，但其核心优势在于为机器",93400,"2026-04-06T19:52:38",[52,14],{"id":62,"github_repo":63,"name":64,"description_en":65,"description_zh":66,"ai_summary_zh":66,"readme_en":67,"readme_zh":68,"quickstart_zh":69,"use_case_zh":70,"hero_image_url":71,"owner_login":72,"owner_name":73,"owner_avatar_url":74,"owner_bio":75,"owner_company":76,"owner_location":76,"owner_email":76,"owner_twitter":76,"owner_website":76,"owner_url":77,"languages":78,"stars":109,"forks":110,"last_commit_at":111,"license":112,"difficulty_score":32,"env_os":113,"env_gpu":114,"env_ram":115,"env_deps":116,"category_tags":122,"github_topics":123,"view_count":32,"oss_zip_url":76,"oss_zip_packed_at":76,"status":17,"created_at":144,"updated_at":145,"faqs":146,"releases":177},7754,"kaito-project\u002Faikit","aikit","🏗️ Fine-tune, build, and deploy open-source LLMs easily!","AIKit 是一个旨在让用户轻松微调、构建和部署开源大语言模型（LLM）的综合平台。它主要解决了本地运行和定制大模型门槛高、依赖复杂环境以及缺乏统一部署标准的痛点，让用户无需 GPU 或联网即可快速上手。\n\n无论是希望探索大模型能力的开发者、需要高效微调模型的研究人员，还是寻求私有化部署方案的企业用户，都能通过 AIKit 获得流畅的体验。其核心亮点在于基于 LocalAI 提供了完全兼容 OpenAI 接口的推理服务，这意味着现有的各类 OpenAI 客户端可直接无缝对接本地模型。此外，AIKit 集成了 Unsloth 以实现快速且低显占用的模型微调，并支持将模型打包为标准的 OCI 制品，便于安全分发和管理。\n\n该平台不仅支持文本对话，还涵盖多模态识别与图像生成功能，兼容 GGUF\u002FGGML 等多种主流模型格式。凭借极简的 Docker 启动方式和完善的 Kubernetes 支持，AIKit 让从个人电脑到边缘计算的模型落地变得简单可靠，同时兼顾了供应链安全与离线环境需求。","# AIKit ✨\n\n\u003Cp align=\"center\">\n\u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002Fkaito-project_aikit_readme_a5cd92470b5a.png\" width=\"200\">\u003Cbr>\n\u003C\u002Fp>\n\nAIKit is a comprehensive platform to quickly get started to host, deploy, build and fine-tune large language models (LLMs).\n\nAIKit offers three main capabilities:\n\n- **Inference**: AIKit uses [LocalAI](https:\u002F\u002Flocalai.io\u002F), which supports a wide range of inference capabilities and formats. LocalAI provides a drop-in replacement REST API that is OpenAI API compatible, so you can use any OpenAI API compatible client, such as [Kubectl AI](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Fkubectl-ai), [Chatbot-UI](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Fchatbot-ui) and many more, to send requests to open LLMs!\n\n- **[Fine-Tuning](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Ffine-tune)**: AIKit offers an extensible fine-tuning interface. It supports [Unsloth](https:\u002F\u002Fgithub.com\u002Funslothai\u002Funsloth) for fast, memory efficient, and easy fine-tuning experience.\n\n- **[OCI Packaging](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fpackaging)**: Package models as OCI artifacts for distribution through any OCI-compliant registry. Supports [CNCF ModelPack](https:\u002F\u002Fgithub.com\u002Fmodelpack\u002Fmodel-spec) specification and generic artifact packaging.\n\n👉 For full documentation, please see [AIKit website](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002F)!\n\n## Features\n\n- 🐳 No GPU, Internet access or additional tools needed except for [Docker](https:\u002F\u002Fdocs.docker.com\u002Fdesktop\u002Finstall\u002Flinux-install\u002F) or [Podman](https:\u002F\u002Fpodman.io)!\n- 🤏 Minimal image size, resulting in less vulnerabilities and smaller attack surface with a custom [chiseled](https:\u002F\u002Fubuntu.com\u002Fcontainers\u002Fchiseled) image\n- 🎵 [Fine-tune support](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Ffine-tune)\n- 📦 [OCI packaging support](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fpackaging) for distributing models as OCI artifacts\n- 🚀 Easy to use declarative configuration for [inference](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fspecs-inference) and [fine-tuning](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fspecs-finetune)\n- ✨ OpenAI API compatible to use with any OpenAI API compatible client\n- 📸 [Multi-modal model support](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fvision)\n- 🖼️ [Image generation support](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fdiffusion)\n- 🦙 Support for GGUF ([`llama`](https:\u002F\u002Fgithub.com\u002Fggerganov\u002Fllama.cpp)) and GGML ([`llama-ggml`](https:\u002F\u002Fgithub.com\u002Fggerganov\u002Fllama.cpp)) models\n- 🚢 [Kubernetes deployment ready](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fkubernetes)\n- 📚 Supports multiple models with a single image\n- 🖥️ Supports [AMD64 and ARM64](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fcreate-images#multi-platform-support) CPUs and [GPU-accelerated inferencing with NVIDIA GPUs](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fgpu)\n- 🔐 Ensure [supply chain security](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fsecurity) with SBOMs, Provenance attestations, and signed images\n- 🌈 Supports air-gapped environments with self-hosted, local, or any remote container registries to store model images for inference on the edge.\n\n## Quick Start\n\nYou can get started with AIKit quickly on your local machine without a GPU!\n\n```bash\ndocker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.1:8b\n```\n\nAfter running this, navigate to [http:\u002F\u002Flocalhost:8080\u002Fchat](http:\u002F\u002Flocalhost:8080\u002Fchat) to access the WebUI!\n\n### API\n\nAIKit provides an OpenAI API compatible endpoint, so you can use any OpenAI API compatible client to send requests to open LLMs!\n\n```bash\ncurl http:\u002F\u002Flocalhost:8080\u002Fv1\u002Fchat\u002Fcompletions -H \"Content-Type: application\u002Fjson\" -d '{\n    \"model\": \"llama-3.1-8b-instruct\",\n    \"messages\": [{\"role\": \"user\", \"content\": \"explain kubernetes in a sentence\"}]\n  }'\n```\n\nOutput should be similar to:\n\n```jsonc\n{\n  \u002F\u002F ...\n    \"model\": \"llama-3.1-8b-instruct\",\n    \"choices\": [\n        {\n            \"index\": 0,\n            \"finish_reason\": \"stop\",\n            \"message\": {\n                \"role\": \"assistant\",\n                \"content\": \"Kubernetes is an open-source container orchestration system that automates the deployment, scaling, and management of applications and services, allowing developers to focus on writing code rather than managing infrastructure.\"\n            }\n        }\n    ],\n  \u002F\u002F ...\n}\n```\n\nThat's it! 🎉 API is OpenAI compatible so this is a drop-in replacement for any OpenAI API compatible client.\n\n## Pre-made Models\n\nAIKit comes with pre-made models that you can use out-of-the-box!\n\nIf it doesn't include a specific model, you can always [create your own images](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fcreate-images), and host in a container registry of your choice!\n\n## CPU\n\n> [!NOTE]\n> AIKit supports both AMD64 and ARM64 CPUs. You can run the same command on either architecture, and Docker will automatically pull the correct image for your CPU.\n>\n> Depending on your CPU capabilities, AIKit will automatically select the most optimized instruction set.\n\n| Model           | Optimization | Parameters | Command                                                                     | Model Name               | License                                                                            |\n| --------------- | ------------ | ---------- | --------------------------------------------------------------------------- | ------------------------ | ---------------------------------------------------------------------------------- |\n| 🦙 Llama 3.2     | Instruct     | 1B         | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.2:1b`   | `llama-3.2-1b-instruct`  | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                        |\n| 🦙 Llama 3.2     | Instruct     | 3B         | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.2:3b`   | `llama-3.2-3b-instruct`  | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                        |\n| 🦙 Llama 3.1     | Instruct     | 8B         | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.1:8b`   | `llama-3.1-8b-instruct`  | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                        |\n| 🦙 Llama 3.3     | Instruct     | 70B        | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.3:70b`  | `llama-3.3-70b-instruct` | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                        |  |\n| Ⓜ️ Mixtral       | Instruct     | 8x7B       | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fmixtral:8x7b`  | `mixtral-8x7b-instruct`  | [Apache](https:\u002F\u002Fchoosealicense.com\u002Flicenses\u002Fapache-2.0\u002F)                          |\n| 🅿️ Phi 4         | Instruct     | 14B        | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fphi4:14b`      | `phi-4-14b-instruct`     | [MIT](https:\u002F\u002Fhuggingface.co\u002Fmicrosoft\u002FPhi-4\u002Fresolve\u002Fmain\u002FLICENSE)                  |\n| 🔡 Gemma 2       | Instruct     | 2B         | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fgemma2:2b`     | `gemma-2-2b-instruct`    | [Gemma](https:\u002F\u002Fai.google.dev\u002Fgemma\u002Fterms)                                         |\n| QwQ             |              | 32B        | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fqwq:32b`       | `qwq-32b`        | [Apache 2.0](https:\u002F\u002Fhuggingface.co\u002FQwen\u002FQwQ-32B\u002Fblob\u002Fmain\u002FLICENSE)        |\n| ⌨️ Codestral 0.1 | Code         | 22B        | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fcodestral:22b` | `codestral-22b`          | [MNLP](https:\u002F\u002Fmistral.ai\u002Flicenses\u002FMNPL-0.1.md)                                    |\n| 🤖 GPT-OSS       |              | 20B        | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fgpt-oss:20b`   | `gpt-oss-20b`            | [Apache 2.0](https:\u002F\u002Fchoosealicense.com\u002Flicenses\u002Fapache-2.0\u002F)                      |\n| 🤖 GPT-OSS       |              | 120B       | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fgpt-oss:120b`  | `gpt-oss-120b`           | [Apache 2.0](https:\u002F\u002Fchoosealicense.com\u002Flicenses\u002Fapache-2.0\u002F)                      |\n\n\n### NVIDIA CUDA\n\n> [!NOTE]\n> To enable GPU acceleration, please see [GPU Acceleration](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fgpu).\n>\n> Please note that only difference between CPU and GPU section is the `--gpus all` flag in the command to enable GPU acceleration.\n\n| Model           | Optimization  | Parameters | Command                                                                                | Model Name               | License                                                                                                                     |\n| --------------- | ------------- | ---------- | -------------------------------------------------------------------------------------- | ------------------------ | --------------------------------------------------------------------------------------------------------------------------- |\n| 🦙 Llama 3.2     | Instruct      | 1B         | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.2:1b`   | `llama-3.2-1b-instruct`  | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                                                                 |\n| 🦙 Llama 3.2     | Instruct      | 3B         | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.2:3b`   | `llama-3.2-3b-instruct`  | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                                                                 |\n| 🦙 Llama 3.1     | Instruct      | 8B         | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.1:8b`   | `llama-3.1-8b-instruct`  | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                                                                 |\n| 🦙 Llama 3.3     | Instruct      | 70B        | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.3:70b`  | `llama-3.3-70b-instruct` | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                                                                 |  |\n| Ⓜ️ Mixtral       | Instruct      | 8x7B       | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fmixtral:8x7b`  | `mixtral-8x7b-instruct`  | [Apache](https:\u002F\u002Fchoosealicense.com\u002Flicenses\u002Fapache-2.0\u002F)                                                                   |\n| 🅿️ Phi 4         | Instruct      | 14B        | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fphi4:14b`      | `phi-4-14b-instruct`     | [MIT](https:\u002F\u002Fhuggingface.co\u002Fmicrosoft\u002FPhi-4\u002Fresolve\u002Fmain\u002FLICENSE)                                                          |\n| 🔡 Gemma 2       | Instruct      | 2B         | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fgemma2:2b`     | `gemma-2-2b-instruct`    | [Gemma](https:\u002F\u002Fai.google.dev\u002Fgemma\u002Fterms)                                                                                  |\n| QwQ             |               | 32B        | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fqwq:32b`       | `qwq-32b`        | [Apache 2.0](https:\u002F\u002Fhuggingface.co\u002FQwen\u002FQwQ-32B\u002Fblob\u002Fmain\u002FLICENSE)                                                 |\n| ⌨️ Codestral 0.1 | Code          | 22B        | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fcodestral:22b` | `codestral-22b`          | [MNLP](https:\u002F\u002Fmistral.ai\u002Flicenses\u002FMNPL-0.1.md)                                                                             |\n| 📸 Flux 1 Dev    | Text to image | 12B        | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fflux1:dev`     | `flux-1-dev`             | [FLUX.1 [dev] Non-Commercial License](https:\u002F\u002Fgithub.com\u002Fblack-forest-labs\u002Fflux\u002Fblob\u002Fmain\u002Fmodel_licenses\u002FLICENSE-FLUX1-dev) |\n| 🤖 GPT-OSS       |               | 20B        | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fgpt-oss:20b`   | `gpt-oss-20b`            | [Apache 2.0](https:\u002F\u002Fchoosealicense.com\u002Flicenses\u002Fapache-2.0\u002F)                                                               |\n| 🤖 GPT-OSS       |               | 120B       | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fgpt-oss:120b`  | `gpt-oss-120b`           | [Apache 2.0](https:\u002F\u002Fchoosealicense.com\u002Flicenses\u002Fapache-2.0\u002F)                                                               |\n\n\n### Apple Silicon (experimental)\n\n> [!NOTE]\n> To enable GPU acceleration on Apple Silicon, please see [Podman Desktop documentation](https:\u002F\u002Fpodman-desktop.io\u002Fdocs\u002Fpodman\u002Fgpu). For more information, please see [GPU Acceleration](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fgpu).\n>\n> Apple Silicon is an _experimental_ runtime and it may change in the future. This runtime is specific to Apple Silicon only, and it will not work as expected on other architectures, including Intel Macs.\n>\n> Only `gguf` models are supported on Apple Silicon.\n\n| Model       | Optimization | Parameters | Command                                                                                                  | Model Name              | License                                                                            |\n| ----------- | ------------ | ---------- | -------------------------------------------------------------------------------------------------------- | ----------------------- | ---------------------------------------------------------------------------------- |\n| 🦙 Llama 3.2 | Instruct     | 1B         | `podman run -d --rm --device \u002Fdev\u002Fdri -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fapplesilicon\u002Fllama3.2:1b` | `llama-3.2-1b-instruct` | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                        |\n| 🦙 Llama 3.2 | Instruct     | 3B         | `podman run -d --rm --device \u002Fdev\u002Fdri -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fapplesilicon\u002Fllama3.2:3b` | `llama-3.2-3b-instruct` | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                        |\n| 🦙 Llama 3.1 | Instruct     | 8B         | `podman run -d --rm --device \u002Fdev\u002Fdri -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fapplesilicon\u002Fllama3.1:8b` | `llama-3.1-8b-instruct` | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                        |\n| 🅿️ Phi 4     | Instruct     | 14B        | `podman run -d --rm --device \u002Fdev\u002Fdri -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fapplesilicon\u002Fphi4:14b`    | `phi-4-14b-instruct`    | [MIT](https:\u002F\u002Fhuggingface.co\u002Fmicrosoft\u002FPhi-4\u002Fresolve\u002Fmain\u002FLICENSE)                  |\n| 🔡 Gemma 2   | Instruct     | 2B         | `podman run -d --rm --device \u002Fdev\u002Fdri -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fapplesilicon\u002Fgemma2:2b`   | `gemma-2-2b-instruct`   | [Gemma](https:\u002F\u002Fai.google.dev\u002Fgemma\u002Fterms)                                         |\n\n## Contributing\n\nWant to contribute to AIKit? Check out our [Contributing Guide](.\u002FCONTRIBUTING.md) for development setup, testing instructions, and contribution guidelines.\n\n## What's next?\n\n👉 For more information and how to fine tune models or create your own images, please see [AIKit website](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002F)!\n","# AIKit ✨\n\n\u003Cp align=\"center\">\n\u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002Fkaito-project_aikit_readme_a5cd92470b5a.png\" width=\"200\">\u003Cbr>\n\u003C\u002Fp>\n\nAIKit 是一个功能全面的平台，可帮助您快速启动、托管、部署、构建和微调大型语言模型（LLMs）。\n\nAIKit 提供三大核心功能：\n\n- **推理**：AIKit 使用 [LocalAI](https:\u002F\u002Flocalai.io\u002F)，它支持广泛的推理能力和多种格式。LocalAI 提供与 OpenAI API 兼容的即插即用式 REST API，因此您可以使用任何兼容 OpenAI API 的客户端，例如 [Kubectl AI](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Fkubectl-ai)、[Chatbot-UI](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Fchatbot-ui) 等，向开源 LLM 发送请求！\n\n- **[微调](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Ffine-tune)**：AIKit 提供可扩展的微调界面。它支持 [Unsloth](https:\u002F\u002Fgithub.com\u002Funslothai\u002Funsloth)，带来快速、内存高效且简便的微调体验。\n\n- **[OCI 打包](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fpackaging)**：将模型打包为 OCI 工件，以便通过任何符合 OCI 标准的注册表进行分发。支持 [CNCF ModelPack](https:\u002F\u002Fgithub.com\u002Fmodelpack\u002Fmodel-spec) 规范以及通用工件打包。\n\n👉 如需完整文档，请访问 [AIKit 官网](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002F)！\n\n## 特性\n\n- 🐳 无需 GPU、互联网连接或额外工具，只需 [Docker](https:\u002F\u002Fdocs.docker.com\u002Fdesktop\u002Finstall\u002Flinux-install\u002F) 或 [Podman](https:\u002F\u002Fpodman.io) 即可！\n- 🤏 镜像体积极小，采用定制的 [chiseled](https:\u002F\u002Fubuntu.com\u002Fcontainers\u002Fchiseled) 镜像，漏洞更少、攻击面更小。\n- 🎵 支持 [微调](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Ffine-tune)\n- 📦 支持 [OCI 打包](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fpackaging)，可将模型作为 OCI 工件分发。\n- 🚀 易于使用的声明式配置，适用于 [推理](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fspecs-inference) 和 [微调](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fspecs-finetune)。\n- ✨ 与 OpenAI API 兼容，可配合任何兼容 OpenAI API 的客户端使用。\n- 📸 支持 [多模态模型](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fvision)。\n- 🖼️ 支持 [图像生成](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fdiffusion)。\n- 🦙 支持 GGUF（`llama`，来自 [gggerganov\u002Fllama.cpp](https:\u002F\u002Fgithub.com\u002Fggerganov\u002Fllama.cpp)）和 GGML（`llama-ggml`，同样来自 [gggerganov\u002Fllama.cpp](https:\u002F\u002Fgithub.com\u002Fggerganov\u002Fllama.cpp)）模型。\n- 🚢 已准备好用于 [Kubernetes 部署](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fkubernetes)。\n- 📚 单一镜像即可支持多种模型。\n- 🖥️ 支持 [AMD64 和 ARM64](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fcreate-images#multi-platform-support) CPU，并可通过 NVIDIA GPU 实现 [GPU 加速推理](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fgpu)。\n- 🔐 通过 SBOM、Provenance 认证和签名镜像，确保 [供应链安全](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fsecurity)。\n- 🌈 支持气隙环境，可使用自托管、本地或任何远程容器注册表来存储模型镜像，以在边缘端进行推理。\n\n## 快速入门\n\n您可以在本地机器上无需 GPU 的情况下快速开始使用 AIKit！\n\n```bash\ndocker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.1:8b\n```\n\n运行后，访问 [http:\u002F\u002Flocalhost:8080\u002Fchat](http:\u002F\u002Flocalhost:8080\u002Fchat)，即可进入 Web UI！\n\n### API\n\nAIKit 提供与 OpenAI API 兼容的端点，因此您可以使用任何兼容 OpenAI API 的客户端向开源 LLM 发送请求！\n\n```bash\ncurl http:\u002F\u002Flocalhost:8080\u002Fv1\u002Fchat\u002Fcompletions -H \"Content-Type: application\u002Fjson\" -d '{\n    \"model\": \"llama-3.1-8b-instruct\",\n    \"messages\": [{\"role\": \"user\", \"content\": \"用一句话解释 Kubernetes\"}]\n  }'\n```\n\n输出应类似于：\n\n```jsonc\n{\n  \u002F\u002F ...\n    \"model\": \"llama-3.1-8b-instruct\",\n    \"choices\": [\n        {\n            \"index\": 0,\n            \"finish_reason\": \"stop\",\n            \"message\": {\n                \"role\": \"assistant\",\n                \"content\": \"Kubernetes 是一个开源的容器编排系统，能够自动化应用程序和服务的部署、扩展与管理，使开发者可以专注于编写代码，而无需操心基础设施的运维。\"\n            }\n        }\n    ],\n  \u002F\u002F ...\n}\n```\n\n就是这样！🎉 该 API 与 OpenAI 兼容，因此可直接替代任何兼容 OpenAI API 的客户端。\n\n## 预制模型\n\nAIKit 自带预制模型，开箱即用！\n\n如果其中未包含您所需的特定模型，您也可以随时 [创建自己的镜像](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fcreate-images)，并将其托管在您选择的容器注册表中！\n\n## CPU\n\n> [!NOTE]\n> AIKit 支持 AMD64 和 ARM64 两种 CPU 架构。您可以在任一架构上运行相同的命令，Docker 会自动拉取适合您 CPU 的正确镜像。\n>\n> 根据您的 CPU 功能，AIKit 会自动选择最优化的指令集。\n\n| 模型           | 优化       | 参数   | 命令                                                                     | 模型名称               | 许可证                                                                            |\n| --------------- | ------------ | ------ | --------------------------------------------------------------------------- | ------------------------ | ---------------------------------------------------------------------------------- |\n| 🦙 Llama 3.2     | 指令       | 1B     | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.2:1b`   | `llama-3.2-1b-instruct`  | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                        |\n| 🦙 Llama 3.2     | 指令       | 3B     | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.2:3b`   | `llama-3.2-3b-instruct`  | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                        |\n| 🦙 Llama 3.1     | 指令       | 8B     | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.1:8b`   | `llama-3.1-8b-instruct`  | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                        |\n| 🦙 Llama 3.3     | 指令       | 70B    | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.3:70b`  | `llama-3.3-70b-instruct` | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                        |  |\n| Ⓜ️ Mixtral       | 指令       | 8x7B   | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fmixtral:8x7b`  | `mixtral-8x7b-instruct`  | [Apache](https:\u002F\u002Fchoosealicense.com\u002Flicenses\u002Fapache-2.0\u002F)                          |\n| 🅿️ Phi 4         | 指令       | 14B    | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fphi4:14b`      | `phi-4-14b-instruct`     | [MIT](https:\u002F\u002Fhuggingface.co\u002Fmicrosoft\u002FPhi-4\u002Fresolve\u002Fmain\u002FLICENSE)                  |\n| 🔡 Gemma 2       | 指令       | 2B     | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fgemma2:2b`     | `gemma-2-2b-instruct`    | [Gemma](https:\u002F\u002Fai.google.dev\u002Fgemma\u002Fterms)                                         |\n| QwQ             |            | 32B    | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fqwq:32b`       | `qwq-32b`        | [Apache 2.0](https:\u002F\u002Fhuggingface.co\u002FQwen\u002FQwQ-32B\u002Fblob\u002Fmain\u002FLICENSE)        |\n| ⌨️ Codestral 0.1 | 代码       | 22B    | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fcodestral:22b` | `codestral-22b`          | [MNLP](https:\u002F\u002Fmistral.ai\u002Flicenses\u002FMNPL-0.1.md)                                    |\n| 🤖 GPT-OSS       |            | 20B    | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fgpt-oss:20b`   | `gpt-oss-20b`            | [Apache 2.0](https:\u002F\u002Fchoosealicense.com\u002Flicenses\u002Fapache-2.0\u002F)                      |\n| 🤖 GPT-OSS       |            | 120B   | `docker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fgpt-oss:120b`  | `gpt-oss-120b`           | [Apache 2.0](https:\u002F\u002Fchoosealicense.com\u002Flicenses\u002Fapache-2.0\u002F)                      |\n\n### NVIDIA CUDA\n\n> [!注意]\n> 要启用 GPU 加速，请参阅 [GPU 加速](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fgpu)。\n>\n> 请注意，CPU 和 GPU 部分之间的唯一区别是用于启用 GPU 加速的命令中的 `--gpus all` 标志。\n\n| 模型           | 优化       | 参数   | 命令                                                                                | 模型名称               | 许可证                                                                                                                     |\n| --------------- | ----------- | ------ | -------------------------------------------------------------------------------------- | ------------------------ | --------------------------------------------------------------------------------------------------------------------------- |\n| 🦙 Llama 3.2     | 指令       | 1B     | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.2:1b`   | `llama-3.2-1b-instruct`  | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                                                                 |\n| 🦙 Llama 3.2     | 指令       | 3B     | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.2:3b`   | `llama-3.2-3b-instruct`  | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                                                                 |\n| 🦙 Llama 3.1     | 指令       | 8B     | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.1:8b`   | `llama-3.1-8b-instruct`  | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                                                                 |\n| 🦙 Llama 3.3     | 指令       | 70B    | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.3:70b`  | `llama-3.3-70b-instruct` | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                                                                 |  |\n| Ⓜ️ Mixtral       | 指令       | 8x7B   | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fmixtral:8x7b`  | `mixtral-8x7b-instruct`  | [Apache](https:\u002F\u002Fchoosealicense.com\u002Flicenses\u002Fapache-2.0\u002F)                                                                   |\n| 🅿️ Phi 4         | 指令       | 14B    | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fphi4:14b`      | `phi-4-14b-instruct`     | [MIT](https:\u002F\u002Fhuggingface.co\u002Fmicrosoft\u002FPhi-4\u002Fresolve\u002Fmain\u002FLICENSE)                                                          |\n| 🔡 Gemma 2       | 指令       | 2B     | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fgemma2:2b`     | `gemma-2-2b-instruct`    | [Gemma](https:\u002F\u002Fai.google.dev\u002Fgemma\u002Fterms)                                                                                  |\n| QwQ             |             | 32B    | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fqwq:32b`       | `qwq-32b`        | [Apache 2.0](https:\u002F\u002Fhuggingface.co\u002FQwen\u002FQwQ-32B\u002Fblob\u002Fmain\u002FLICENSE)                                                 |\n| ⌨️ Codestral 0.1 | 代码       | 22B    | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fcodestral:22b` | `codestral-22b`          | [MNLP](https:\u002F\u002Fmistral.ai\u002Flicenses\u002FMNPL-0.1.md)                                                                             |\n| 📸 Flux 1 Dev    | 文本到图像 | 12B    | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fflux1:dev`     | `flux-1-dev`             | [FLUX.1 [dev] 非商业许可](https:\u002F\u002Fgithub.com\u002Fblack-forest-labs\u002Fflux\u002Fblob\u002Fmain\u002Fmodel_licenses\u002FLICENSE-FLUX1-dev) |\n| 🤖 GPT-OSS       |             | 20B    | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fgpt-oss:20b`   | `gpt-oss-20b`            | [Apache 2.0](https:\u002F\u002Fchoosealicense.com\u002Flicenses\u002Fapache-2.0\u002F)                                                               |\n| 🤖 GPT-OSS       |             | 120B   | `docker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fgpt-oss:120b`  | `gpt-oss-120b`           | [Apache 2.0](https:\u002F\u002Fchoosealicense.com\u002Flicenses\u002Fapache-2.0\u002F)                                                               |\n\n\n### Apple Silicon（实验性）\n\n> [!注意]\n> 要在 Apple Silicon 上启用 GPU 加速，请参阅 [Podman Desktop 文档](https:\u002F\u002Fpodman-desktop.io\u002Fdocs\u002Fpodman\u002Fgpu)。有关更多信息，请参阅 [GPU 加速](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fgpu)。\n>\n> Apple Silicon 是一种 _实验性_ 运行时环境，未来可能会发生变化。此运行时环境仅适用于 Apple Silicon，无法在其他架构上正常工作，包括 Intel Mac。\n>\n> Apple Silicon 仅支持 `gguf` 模型。\n\n| 模型       | 优化       | 参数   | 命令                                                                                                  | 模型名称              | 许可证                                                                            |\n| ----------- | ----------- | ------ | -------------------------------------------------------------------------------------------------------- | ----------------------- | ---------------------------------------------------------------------------------- |\n| 🦙 Llama 3.2 | 指令       | 1B     | `podman run -d --rm --device \u002Fdev\u002Fdri -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fapplesilicon\u002Fllama3.2:1b` | `llama-3.2-1b-instruct` | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                        |\n| 🦙 Llama 3.2 | 指令       | 3B     | `podman run -d --rm --device \u002Fdev\u002Fdri -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fapplesilicon\u002Fllama3.2:3b` | `llama-3.2-3b-instruct` | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                        |\n| 🦙 Llama 3.1 | 指令       | 8B     | `podman run -d --rm --device \u002Fdev\u002Fdri -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fapplesilicon\u002Fllama3.1:8b` | `llama-3.1-8b-instruct` | [Llama](https:\u002F\u002Fai.meta.com\u002Fllama\u002Flicense\u002F)                                        |\n| 🅿️ Phi 4     | 指令       | 14B    | `podman run -d --rm --device \u002Fdev\u002Fdri -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fapplesilicon\u002Fphi4:14b`    | `phi-4-14b-instruct`    | [MIT](https:\u002F\u002Fhuggingface.co\u002Fmicrosoft\u002FPhi-4\u002Fresolve\u002Fmain\u002FLICENSE)                  |\n| 🔡 Gemma 2   | 指令       | 2B     | `podman run -d --rm --device \u002Fdev\u002Fdri -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fapplesilicon\u002Fgemma2:2b`   | `gemma-2-2b-instruct`   | [Gemma](https:\u002F\u002Fai.google.dev\u002Fgemma\u002Fterms)                                         |\n\n## 贡献\n\n想为 AIKit 做贡献吗？请查看我们的 [贡献指南](.\u002FCONTRIBUTING.md)，了解开发设置、测试说明和贡献规范。\n\n## 下一步是什么？\n\n👉 欲了解更多关于如何微调模型或创建自有图像的信息，请访问 [AIKit 官网](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002F)！","# AIKit 快速上手指南\n\nAIKit 是一个综合平台，旨在帮助用户快速托管、部署、构建和微调大型语言模型（LLM）。它基于 LocalAI 提供兼容 OpenAI API 的推理服务，支持无需 GPU 即可在本地运行多种主流模型。\n\n## 环境准备\n\n*   **操作系统**：支持 Linux、macOS 和 Windows（需安装 Docker Desktop）。\n*   **架构支持**：自动适配 AMD64 (x86_64) 和 ARM64 (如 Apple Silicon, Raspberry Pi) 架构。\n*   **前置依赖**：\n    *   必须安装 [Docker](https:\u002F\u002Fdocs.docker.com\u002Fdesktop\u002Finstall\u002Flinux-install\u002F) 或 [Podman](https:\u002F\u002Fpodman.io\u002F)。\n    *   **可选**：若需 GPU 加速推理，需安装 NVIDIA 驱动及 Docker GPU 运行时支持。\n*   **网络要求**：首次运行需联网拉取镜像，后续支持离线（Air-gapped）环境运行。\n\n> **注意**：AIKit 无需额外安装 Python 环境、CUDA 工具包或其他 AI 框架，所有依赖已封装在容器镜像中。\n\n## 安装步骤\n\nAIKit 采用容器化交付，无需传统“安装”过程，只需拉取并运行对应的 Docker 镜像即可。\n\n### 1. 基础运行（CPU 模式）\n无需 GPU，直接在本地启动一个 Llama 3.1 8B 模型服务：\n\n```bash\ndocker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.1:8b\n```\n\n### 2. GPU 加速运行（NVIDIA CUDA）\n若主机配备 NVIDIA GPU，只需添加 `--gpus all` 参数即可启用硬件加速：\n\n```bash\ndocker run -d --rm --gpus all -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fllama3.1:8b\n```\n\n> **提示**：Docker 会自动根据当前 CPU 架构（AMD64\u002FARM64）拉取匹配的镜像版本。\n\n## 基本使用\n\n启动容器后，AIKit 将提供 Web 界面和兼容 OpenAI 标准的 API 接口。\n\n### 1. 访问 Web 界面\n在浏览器中打开以下地址，即可通过图形化界面与模型对话：\n\n```text\nhttp:\u002F\u002Flocalhost:8080\u002Fchat\n```\n\n### 2. 调用 API\nAIKit 完全兼容 OpenAI API 格式，可直接使用任何支持 OpenAI 协议的客户端（如 Chatbot-UI、Kubectl AI 等）或发送 HTTP 请求。\n\n**示例命令：**\n\n```bash\ncurl http:\u002F\u002Flocalhost:8080\u002Fv1\u002Fchat\u002Fcompletions -H \"Content-Type: application\u002Fjson\" -d '{\n    \"model\": \"llama-3.1-8b-instruct\",\n    \"messages\": [{\"role\": \"user\", \"content\": \"explain kubernetes in a sentence\"}]\n  }'\n```\n\n**预期返回示例：**\n\n```jsonc\n{\n  \u002F\u002F ...\n    \"model\": \"llama-3.1-8b-instruct\",\n    \"choices\": [\n        {\n            \"index\": 0,\n            \"finish_reason\": \"stop\",\n            \"message\": {\n                \"role\": \"assistant\",\n                \"content\": \"Kubernetes is an open-source container orchestration system that automates the deployment, scaling, and management of applications and services, allowing developers to focus on writing code rather than managing infrastructure.\"\n            }\n        }\n    ],\n  \u002F\u002F ...\n}\n```\n\n### 3. 更多预置模型\n除了 Llama 3.1，AIKit 还支持一键运行 Mixtral, Phi-4, Gemma, Codestral 等多种模型。只需替换镜像标签即可，例如运行 Phi-4：\n\n```bash\ndocker run -d --rm -p 8080:8080 ghcr.io\u002Fkaito-project\u002Faikit\u002Fphi4:14b\n```\n\n详细模型列表及对应命令请参考官方文档或项目 README 中的\"Pre-made Models\"章节。","某初创公司的算法团队需要在离线环境中快速部署并微调一个定制化的客服大模型，以适配内部知识库。\n\n### 没有 aikit 时\n- **环境配置繁琐**：团队成员需手动安装 CUDA、PyTorch 及各类依赖库，常因版本冲突导致数天的环境调试时间浪费。\n- **微调门槛极高**：缺乏显存优化方案，在有限硬件上微调模型极易显存溢出（OOM），且代码改造复杂，难以快速迭代。\n- **交付与集成困难**：模型打包格式不统一，无法直接通过标准容器 registry 分发；同时需额外开发适配层才能对接现有的 OpenAI 兼容客户端。\n- **离线部署受阻**：在无网环境下，缺乏将模型预打包为独立镜像的机制，导致边缘端部署几乎不可行。\n\n### 使用 aikit 后\n- **一键启动服务**：仅需一条 Docker 命令即可拉起包含推理引擎的完整服务，无需安装 GPU 驱动或额外工具，即刻通过 WebUI 或 API 访问。\n- **高效低耗微调**：内置 Unsloth 支持，利用声明式配置即可启动显存高效的微调任务，大幅降低硬件要求并缩短训练周期。\n- **标准化分发与兼容**：自动将模型封装为 OCI 制品推送到私有仓库，原生提供 OpenAI 兼容接口，现有业务代码零修改即可切换至本地模型。\n- **无缝离线运行**：支持将模型与运行时一起打包成最小化安全镜像，轻松在有防火墙或完全隔离的内网环境中稳定运行。\n\naikit 通过屏蔽底层基础设施复杂度，让团队能专注于模型业务逻辑，实现了从“配置环境”到“应用模型”的效率飞跃。","https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002Fkaito-project_aikit_9c181e5b.png","kaito-project","Kaito","https:\u002F\u002Foss.gittoolsai.com\u002Favatars\u002Fkaito-project_28046849.png","Kubernetes AI Toolchain Operator",null,"https:\u002F\u002Fgithub.com\u002Fkaito-project",[79,83,86,90,94,98,101,105],{"name":80,"color":81,"percentage":82},"Go","#00ADD8",89.7,{"name":84,"color":85,"percentage":10},"JavaScript","#f1e05a",{"name":87,"color":88,"percentage":89},"Python","#3572A5",2.1,{"name":91,"color":92,"percentage":93},"Makefile","#427819",1.6,{"name":95,"color":96,"percentage":97},"CSS","#663399",1.3,{"name":99,"color":81,"percentage":100},"Go Template",1,{"name":102,"color":103,"percentage":104},"Shell","#89e051",0.9,{"name":106,"color":107,"percentage":108},"Dockerfile","#384d54",0.4,515,59,"2026-04-13T04:01:24","MIT","Linux, macOS, Windows","非必需。支持 NVIDIA GPU 加速（需安装 NVIDIA Container Toolkit 并使用 --gpus all 标志），未明确具体型号和显存要求，取决于运行的模型大小；也支持纯 CPU 运行（AMD64\u002FARM64）。","未说明（取决于所选模型大小，例如运行 70B 模型通常需要大量内存）",{"notes":117,"python":118,"dependencies":119},"该工具基于容器化部署，无需直接安装 Python 环境或深度学习库。核心依赖是 Docker 或 Podman。支持 AMD64 和 ARM64 架构。在无 GPU 环境下可直接运行，有 NVIDIA GPU 时可通过添加启动参数启用加速。支持离线（Air-gapped）环境部署。","未说明",[120,121],"Docker","Podman",[14,35,15,13],[124,125,126,127,128,129,130,131,132,133,134,135,136,137,138,139,140,141,142,143],"buildkit","large-language-models","llm","docker","inference","kubernetes","nvidia","ai","llama","chatgpt","localllama","open-source-llm","mistral","mixtral","gpt","openai","fine-tuning","finetuning","gemma","open-llm","2026-03-27T02:49:30.150509","2026-04-16T01:44:49.378357",[147,152,157,162,167,172],{"id":148,"question_zh":149,"answer_zh":150,"source_url":151},34715,"如何添加新的预构建模型（如代码补全模型或 Phi-3）？","维护者通常会直接将流行的新模型（如 Codestral、Phi-3 mini 4k 版本）添加到预构建模型列表中。您可以访问官方文档查看最新支持的模型列表：https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fdocs\u002Fpremade-models。如果您需要的模型尚未支持，可以参考文档自行创建镜像：https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fcreate-images。对于特定版本（如 Phi-3 128k），可能需要等待底层依赖（如 llama.cpp）支持后由社区贡献。","https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fissues\u002F236",{"id":153,"question_zh":154,"answer_zh":155,"source_url":156},34716,"遇到 'could not load model: rpc error: code = Unavailable desc = error reading from server: EOF' 错误怎么办？","该错误通常源于底层的 LocalAI 组件而非 AIKit 本身。如果在本地复现了该问题，建议直接前往 LocalAI 仓库提交 Issue 或查找现有解决方案（例如检查 CPU 标志位 AVX 是否启用）。一旦 LocalAI 发布修复版本，AIKit 在更新后会自动包含该修复。相关讨论可参考：https:\u002F\u002Fgithub.com\u002Fmudler\u002FLocalAI\u002Fissues\u002F1453","https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fissues\u002F41",{"id":158,"question_zh":159,"answer_zh":160,"source_url":161},34717,"如何基于自定义 GGUF 模型文件创建本地服务镜像？","请参考官方文档中关于创建新镜像的指南：https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fcreate-images。文档中包含了详细的步骤说明。如果在操作过程中遇到类似“无法将镜像复制到 docker 格式”的错误，请确保环境配置正确，并在必要时开启新 Issue 提供详细的环境信息和日志以便排查。","https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fissues\u002F123",{"id":163,"question_zh":164,"answer_zh":165,"source_url":166},34718,"为什么某些大上下文模型（如 Phi-3 128k）暂时无法使用？","这通常是因为底层的推理引擎（如 llama.cpp）尚未完全支持特定的缩放类型（如 longrope scaling）。例如，Phi-3 128k 版本曾因 llama.cpp 的支持问题而暂时不可用。目前通常先提供标准上下文版本（如 4k 版本）。您可以关注 llama.cpp 的更新进度，待其支持后，AIKit 也会随之更新支持。","https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fissues\u002F209",{"id":168,"question_zh":169,"answer_zh":170,"source_url":171},34719,"LocalAI 在处理函数调用（Function Calling）时发生 Panic 崩溃如何解决？","此类崩溃通常是由于 LocalAI 上游版本的 Bug 导致的。解决方案是等待并升级到已合并修复的新版本 LocalAI。维护者会跟踪上游修复进度，一旦新版本发布，AIKit 将会进行升级以包含该修复。","https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fissues\u002F638",{"id":173,"question_zh":174,"answer_zh":175,"source_url":176},34720,"更新工作流或工具后出现错误，该如何处理？","如果是由于 GitHub Actions 工作流更新导致的问题，尝试重新固定（repin）actions 的版本通常可以解决问题。可以使用命令 `gh aw add githubnext\u002Fagentics\u002Fworkflows\u002Fissue-triage.md@\u003Ccommit-hash>` 来指定特定版本。更多用法指南请参考：https:\u002F\u002Fgithubnext.github.io\u002Fgh-aw\u002Ftools\u002Fcli\u002F","https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fissues\u002F668",[178,183,188,193,198,203,208,213,218,223,228,233,238,243,248,253,258,263,268,273],{"id":179,"version":180,"summary_zh":181,"released_at":182},272070,"v0.21.0","## 功能\n- 将 LocalAI 更新至 v4.0.0 (#750) [#750](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F750) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fb35126dd170c1cbc4a5cdee0cd88ce3c18442c5f))\n\n## 错误修复\n- 为文件名添加 org.opencontainers.image.title (#681) [#681](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F681) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F9309553bac53a10cf599fded2bdeed93b9471cdd))\n- 打包工具排除项 (#701) [#701](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F701) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F4109be199c583f1d26d7e669a1f4ed51c1ec0bbb))\n- 为 Llama 模型启用语法约束的函数调用 (#711) [#711](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F711) ([Robbie Cronin](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F88649d5bea2694c4e52c7a84c90f3d66344c5e9c))\n\n## 文档\n- 修复打包工具文档 (#680) [#680](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F680) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F06e99f3ce372ca26e957fcd93691f9a285e51bc2))\n- 用基于研究的 CLAUDE.md 替代 AGENTS.md (#749) [#749](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F749) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fe331af307ab880b64c87a471aa0258b3026a568e))\n\n## 杂项任务\n- 准备 v0.20.0 版本发布 (#679) [#679](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F679) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fd99dcec93c7c21dc11801f582827156171d9a441))\n- **依赖**: 将 Python 从 3.12-slim 升级至 3.14-slim [#684](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F684) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F116bf7ac350cd4cf5efda6bbe62fef2b1ef042b2))\n- **依赖**: 将 Python 从 3.12-slim 升级至 3.14-slim (#684) [#684](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F684) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fda83c62a8dc056d043cb32f41be7897099661436))\n- **依赖**: 升级 github.com\u002Fcontainerd\u002Fcontainerd\u002Fv2 [#687](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F687) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Ff2b89973909ad138107748b37d00f3ee339e073b))\n- **依赖**: 在 go_modules 组中，将 github.com\u002Fcontainerd\u002Fcontainerd\u002Fv2 从 2.1.4 升级至 2.1.5，涉及 1 个目录 (#687) [#687](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F687) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F7cfe1de9c9087dc36c6e771feabc3b12bd302fc5))\n- **依赖**: 升级 github.com\u002Fcontainerd\u002Fplatforms [#688](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F688) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F54f2392cb7da9460e46cd39f2b4062bd7f74a1f2))\n- **依赖**: 将 github.com\u002Fcontainerd\u002Fplatforms 从 1.0.0-rc.1 升级至 1.0.0-rc.2 (#688) [#688](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F688) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fc23b9066b141a587d06e0f57d0b3289b0da0e54f))\n- **依赖**: 将 github.com\u002Fmoby\u002Fbuildkit 从 0.25.1 升级至 0.25.2 [#690](https:\u002F\u002Fgithub…","2026-03-17T22:08:36",{"id":184,"version":185,"summary_zh":186,"released_at":187},272071,"v0.20.3","# 重要变更\n- 📦 [将 AI 模型打包为 OCI 工件](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fpackaging)，以便通过容器注册表从本地、HTTP 或 Hugging Face 源进行分发，同时支持通用 OCI 格式和 [CNCF ModelPack](https:\u002F\u002Fgithub.com\u002Fmodelpack\u002Fmodel-spec) 规范。\n- ✨ 更新至 LocalAI [v3.8.0](https:\u002F\u002Fgithub.com\u002Fmudler\u002FLocalAI\u002Freleases\u002Ftag\u002Fv3.8.0)。\n\n## 杂项任务\n- 合并 #704 (#705) [#705](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F705) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F7f5da5c89db115e217ee9947e0ced49603c997b8))\n- 准备 v0.20.3 版本发布 (#706) [#706](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F706) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F3493ad622b64fe1f9b62b8c787b5fee64947632f))","2025-12-04T04:48:16",{"id":189,"version":190,"summary_zh":191,"released_at":192},272072,"v0.20.2","# 重要变更\n- 📦 [将 AI 模型打包为 OCI 工件](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fpackaging)，以便通过容器注册表从本地、HTTP 或 Hugging Face 源进行分发，同时支持通用 OCI 格式和 [CNCF ModelPack](https:\u002F\u002Fgithub.com\u002Fmodelpack\u002Fmodel-spec) 规范。\n- ✨ 更新至 LocalAI [v3.7.0](https:\u002F\u002Fgithub.com\u002Fmudler\u002FLocalAI\u002Freleases\u002Ftag\u002Fv3.7.0)。\n\n## 杂项任务\n- 准备 v0.20.2 版本发布 (#703) [#703](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F703) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F346893ebd7875a63bce793397009ab7a4bfbac8d))\n\n## 提交记录\n- 76ba70a：挑选 #693 和 #701 的提交 (#702)（Sertaç Özercan）[#702](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F702)","2025-12-03T00:54:41",{"id":194,"version":195,"summary_zh":196,"released_at":197},272073,"v0.20.1","# 重要变更\n📦 将 AI 模型打包为 OCI 工件（[文档](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fpackaging)），以便通过容器注册表从本地、HTTP 或 Hugging Face 源进行分发，同时支持通用 OCI 格式和 [CNCF ModelPack](https:\u002F\u002Fgithub.com\u002Fmodelpack\u002Fmodel-spec) 规范。\n\n## 错误修复\n- 合并 #681 (#682) [#682](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F682) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Faacc63a45b2d6b279f822626ff7a5bbed498e619))\n\n## 日常维护\n- 准备 v0.20.1 版本发布 (#683) [#683](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F683) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F52f6459f459316d67b65d12af129c9993aad80d5))","2025-10-30T21:43:44",{"id":199,"version":200,"summary_zh":201,"released_at":202},272074,"v0.20.0","# 重要变更\n📦 将 AI 模型打包为 OCI 工件（[Package AI models as OCI artifacts](https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fpackaging)），以便通过容器注册表从本地、HTTP 或 Hugging Face 源进行分发，同时支持通用 OCI 格式和 [CNCF ModelPack](https:\u002F\u002Fgithub.com\u002Fmodelpack\u002Fmodel-spec) 规范。\n\n# 哪些内容发生了变化？\n\n## 功能\n- OCI 打包器 (#649) [#649](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F649) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fa2648c208789c0ed0087e44b6bf727637e2055bf))\n\n## Bug 修复       \n- 处理多元后端及 CUDA 的 CPU 回退 (#618) [#618](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F618) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F018e1c72f9304289a4e6f4c5c3257ff1223f4333))\n\n## 持续集成\n- 使用大型运行器 (#619) [#619](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F619) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fbb673422b9c52a8c22d7b425367566a711d69e6e))\n- 在大型运行器上更新模型 (#626) [#626](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F626) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F8f25a6b71ba33aee916a549b8710bf77b095dfde))\n- 将 localai 镜像到 ghcr (#644) [#644](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F644) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fee3b556d9b6787e22ae3b4d43a400e949be19ba1))\n- 锁定问题分类操作 (#669) [#669](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F669) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F104c79ad19abba7ad418b9369ff1ce8ada6aebaa))\n- 锁定问题分类操作 (#670) [#670](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F670) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fa6dc17c8ffee0a49f0769613a6e210565ba9da24))\n- 锁定问题分类操作 第三部分 (#671) [#671](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F671) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fe008514ba91f9d37d2025a6ddeb34c5d2bfd6e98))\n- 更新 gh aw (#672) [#672](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F672) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F85d3b6a90bb5d2a15d5a9ae0f1b85bd0a320f611))\n\n## 杂项任务\n- 准备 v0.19.0 发布 (#610) [#610](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F610) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F1199e5b0ce89ff2cd7f49029d0fdeefd4e797d18))\n- **依赖**: 将 actions-all 组更新两次 [#615](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F615) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fcc8c519f299104a39104a4cbe89a42f5bdef96ae))\n- **依赖**: 将 actions-all 组更新两次 (#615) [#615](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F615) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fc73b923bd87f181c3cf7cfa7c903a778c8afebdf))\n- **依赖**: 将 google.golang.org\u002Fgrpc 从 1.74.2 升级到 1.75.0 (#617) [#617](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F617) ([dependabot[bot]](htt","2025-10-27T19:33:52",{"id":204,"version":205,"summary_zh":206,"released_at":207},272075,"v0.19.4","## 杂项任务\n- 挑选 #654 及其相关提交 (#659) [#659](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F659) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fb7ae2cccf6f09712358074d898c4464245b7b844))\n- 准备 v0.19.4 版本发布 (#660) [#660](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F660) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F2f89313a2ceca5af76c8eccec93adba1ae3b1ba7))","2025-10-10T20:26:49",{"id":209,"version":210,"summary_zh":211,"released_at":212},272076,"v0.19.3","## 持续集成\n- 将 localai 镜像到 ghcr (#645) [#645](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F645) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fe203d07bd3a0966f5162f4ce8b0d9252928087b6))\n\n## 杂项任务\n- 准备 v0.19.3 版本发布 (#646) [#646](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F646) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F42e4dae18a54212863a39593daadef7945e63d44))","2025-09-11T04:56:21",{"id":214,"version":215,"summary_zh":216,"released_at":217},272077,"v0.19.2","## 错误修复\n- 合并 #629 (#631) [#631](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F631) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Ff93a979f582e873320b39cae2f2592ab8bd249e8))\n\n## 杂项任务\n- 准备 v0.19.2 版本发布 (#632) [#632](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F632) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fa57b8216ccc732693278367b36e86c8e39131cff))","2025-08-29T16:22:42",{"id":219,"version":220,"summary_zh":221,"released_at":222},272078,"v0.19.1","## 错误修复\n- 精选 #618 和 #619 (#620) [#620](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F620) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F039cbfd41af85ea9906572fd0599339daba0bf73))\n\n## 杂项任务\n- 准备 v0.19.0 版本发布 (#609) (#624) [#624](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F624) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fd92c5bfdd34a213dd5e014f572841a63e3f94e38))\n- 准备 v0.19.1 版本发布 (#625) [#625](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F625) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F9902f877c620547c537dcacca467297f2d6fab46))","2025-08-27T02:29:52",{"id":224,"version":225,"summary_zh":226,"released_at":227},272079,"v0.19.0","## 重要变更\n- ✨ 更新至 LocalAI [v3.4.0](https:\u002F\u002Fgithub.com\u002Fmudler\u002FLocalAI\u002Freleases\u002Ftag\u002Fv3.4.0)。\n- 🤖 在[预置模型](https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fdocs\u002Fpremade-models)中新增 GPT-OSS 20B 和 120B。\n- 🅿️ 在[预置模型](https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fdocs\u002Fpremade-models)中新增 Phi 4。\n- 🥺 将 QWQ 32B 更新为非预览版，纳入[预置模型](https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fdocs\u002Fpremade-models)。\n- 💀 下线 Phi 3.5 和 QWQ 32B 的预览版本。\n\n\u003Cdetails>\n\n\u003Csummary>完整变更列表\u003C\u002Fsummary>\n\n## 功能\n- gpt-oss 20b 和 120b (#599) [#599](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F599) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F95ecd01ad5d2060b06e08acb1f11517bcfcb763a))\n- 新增 Phi4 并将 QWQ 更新为非预览版 (#601) [#601](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F601) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F2d09f51456da78027af177a37008e4c266695c7a))\n\n## Bug 修复\n- grpcio 版本问题 (#550) [#550](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F550) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Ff75cb73147d964ee8446ad0259c5c3d9258f05b7))\n\n## 文档\n- 更新 Flux 示例 (#587) [#587](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F587) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F77942a0ca8ae41c1ba14b9d3a5f524e55674a86e))\n\n## 测试\n- 测试所有分支 (#608) [#608](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F608) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F0fe7bdb346c192bcf2599394b58e2256d8731e52))\n\n## 持续集成\n- GPU 测试输入 (#554) [#554](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F554) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F76e159ca054f5355130bd2f13cce7f20b6ed6a06))\n- 暂时禁用 Helm 测试 (#551) [#551](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F551) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fb7d477c9f8ef54e0a1d4f2078808a81a24781cb6))\n- 将 *.githubusercontent.com 加入白名单 (#566) [#566](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F566) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fe50f806c7c19300ccf51d67a0478603ab090c8d6))\n- 将工作流权限更新至作业级别 (#570) [#570](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F570) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002Fa9f32636c4ba3b2c184f243659cc4506666c7e73))\n- 启用 CI 测试在发布分支上运行 (#581) [#581](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F581) ([Copilot](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F18b040963caaa8d5eadd1122eb259d3308ffed6a))\n\n## 杂项\n- 准备 v0.18.0 发布 (#496) [#496](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F496) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F921460d09e2ba5df0a8a2f36e0d26212f5533a17))\n- **依赖**: 将 Go 从 `6260304` 升级至 `b970e6d` [#497](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F497) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F6009b","2025-08-21T19:11:20",{"id":229,"version":230,"summary_zh":231,"released_at":232},272080,"v0.18.2","## Announcement\r\n\r\n🎉 First release from https:\u002F\u002Fgithub.com\u002Fkaito-project (CNCF sandbox) organization!\r\n\r\n- New website URL: https:\u002F\u002Fkaito-project.github.io\u002Faikit\r\n\r\n- New charts URL: https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fcharts\r\n`helm repo add aikit https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fcharts --force-update` to migrate existing Helm repo\r\n\r\n- AIKit BuildKit frontend image is moved to [`ghcr.io\u002Fkaito-project\u002Faikit\u002Faikit:latest`](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpkgs\u002Fcontainer\u002Faikit%2Faikit)\r\n\r\n- New pre-made image registries are moved under `ghcr.io\u002Fkaito-project\u002Faikit\u002F` : https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fdocs\u002Fpremade-models\r\n\r\n## Chores\r\n- Prepare v0.18.2 release (#574) [#574](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F574) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F639e2bbcddf8e5fa4dabfed1ce0825c0721793b1))\r\n- cherry pick kaito org migration (#575) [#575](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fpull\u002F575) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fkaito-project\u002Faikit\u002Fcommit\u002F46a547344dd62e0ac495332815358d5227df00da))","2025-08-07T18:12:29",{"id":234,"version":235,"summary_zh":236,"released_at":237},272081,"v0.18.1","## Breaking Changes\r\n⚠️ This will be the last release before AIKit moves to https:\u002F\u002Fgithub.com\u002Fkaito-project (CNCF sandbox project) organization. \r\n- Existing pre-made images will continue to work, but no new images will be published under `ghcr.io\u002Fsozercan`.\r\n- New images will be published under `ghcr.io\u002Fkaito-project\u002Faikit\u002F`\r\n- Documentation website will be `https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002F`\r\n- Helm repo will need to be updated to `https:\u002F\u002Fkaito-project.github.io\u002Faikit\u002Fcharts`\r\n\r\n## Chores\r\n- cherry pick chores into release 0.18 (#564) [#564](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F564) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fc68be4bc7228d5622029249c04ee82e3a897e987))\r\n- Prepare v0.18.1 release (#565) [#565](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F565) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F1081cdfe03ab60d39300aa27c1bd86007c0e3bb1))\r\n- Prepare v0.18.0 release (#495) (#567) [#567](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F567) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F7d7b070d2151fbb97bf5c8e4b48f77fa9f570d43))","2025-08-07T02:19:15",{"id":239,"version":240,"summary_zh":241,"released_at":242},272082,"v0.18.0","## Continuous Integration\n- bump golang 1.24 (#489) [#489](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F489) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F4737ace3a6e74ff9592381a6ea1ff6b8f2c4c78e))\n\n## Chores\n- Prepare v0.17.0 release (#490) [#490](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F490) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F1c40cbd75328a94759032ef6b922cbb455b65e3b))\n- **deps**: Bump the actions-all group with 6 updates [#493](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F493) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fe7fd73885f8d468b61d36f13c24d3831108ea59b))\n- **deps**: Bump the actions-all group with 6 updates (#493) [#493](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F493) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F9d8a7b52f0ecef2bb9fb40dbc90caa5a34ed94ec))\n- **deps**: Bump github.com\u002Fmoby\u002Fbuildkit from 0.19.0 to 0.20.0 [#494](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F494) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F6dbfe8d2a6cb214292fcf9bf85331e93c5143b23))\n- **deps**: Bump github.com\u002Fmoby\u002Fbuildkit from 0.19.0 to 0.20.0 (#494) [#494](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F494) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fa1624c3c22fcd217417ea460f8acd09ec75ff545))\n- bump localai v2.26.0 (#487) [#487](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F487) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F921e04a83cd8bf9a5da754991e2d65d5bdcd8c22))\n- Prepare v0.18.0 release (#495) [#495](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F495) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F2b433d4a089cbd35cc412ae5d787b27c981f234c))","2025-02-26T06:24:44",{"id":244,"version":245,"summary_zh":246,"released_at":247},272083,"v0.17.0","## Continuous Integration\n- auto merge dependabot prs (#477) [#477](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F477) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Ff595f2db550f290f10975dba3425b6877b10ca5f))\n\n## Chores\n- Prepare v0.16.0 release (#456) [#456](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F456) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F009060103fb832adbaca22a182bbe06f630cad9f))\n- **deps**: Bump golang.org\u002Fx\u002Fcrypto from 0.27.0 to 0.31.0 in the go_modules group (#457) [#457](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F457) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fbd3a6013494dc1e06aa848aa1710ac264d27333b))\n- **deps**: Bump nanoid from 3.3.7 to 3.3.8 in \u002Fwebsite in the npm_and_yarn group across 1 directory (#458) [#458](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F458) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fe96f6a9600fcf6600631d9502fc3009c2da6a7e2))\n- **deps**: Bump github.com\u002Fmoby\u002Fbuildkit from 0.18.1 to 0.18.2 (#463) [#463](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F463) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fe3fb25e8c43bfb299b8589f8f80168cb789115c4))\n- **deps**: Bump golang from `ef30001` to `2e83858` (#465) [#465](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F465) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Ff0437eaa487f9b94234e202a971918c098b73080))\n- **deps**: Bump kubectl from v1.31.3 to v1.32.1 in \u002Fcharts\u002Faikit (#469) [#469](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F469) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F91cff3a72cf2e9ed9c0543c613adaf847c9f26aa))\n- **deps**: Bump the actions-all group across 1 directory with 10 updates (#470) [#470](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F470) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F87820ac7ddeba5091fc3386416220f752d46c394))\n- **deps**: Bump golang.org\u002Fx\u002Fnet from 0.29.0 to 0.33.0 in the go_modules group (#471) [#471](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F471) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F2212bd71b5c1e095c54ccd52ebbd9190e8ba0683))\n- **deps**: Bump google.golang.org\u002Fgrpc from 1.68.1 to 1.70.0 (#472) [#472](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F472) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Ff9b5b4fa903e8c220752582cf0a519ee602f7653))\n- **deps**: Bump golang from `2e83858` to `3149bc5` (#476) [#476](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F476) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F6e9a02f57811890f9fc29ea2673aedb26e0af2f4))\n- **deps**: Bump the actions-all group with 3 updates (#475) [#475](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F475) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F2b499b6295d50c4f82537942f2f59c6876a266e6))\n- **deps**: Bump github.com\u002Fmoby\u002Fbuildkit from 0.18.2 to 0.19.0 (#474) [#474](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F474) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F2182ae9402b81c38e87e0e784fcaf8b3fa179f75))\n- update localai to v2.25.0 (#473) [#473](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F473) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fcf0d802b188ca078cc2466a29411200209e0e4d0))\n- **deps**: Bump the actions-all group across 1 directory with 6 updates [#480](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F480) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F081ea96c6a3344219ecf88b6255f5fc2462f62c2))\n- **deps**: Bump golang from `3149bc5` to `441f59f` [#481](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F481) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F20d343129bf5802a098f45c1d8a639de1ff68d9b))\n- **deps**: Bump golang.org\u002Fx\u002Fsync from 0.10.0 to 0.11.0 [#482](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F482) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F06d05d932eb57797725a4974e4af1e56a0d890c6))\n- **deps**: Bump golang.org\u002Fx\u002Fsync from 0.10.0 to 0.11.0 (#482) [#482](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F482) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F80fd1fef86f120a204ebe553bcae9b4207db4a84))\n- **deps**: Bump golang from `3149bc5` to `441f59f` (#481) [#481](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F481) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F4337f6762f64592d76ac7b2d8d912cebae823cbe))\n- **deps**: Bump the actions-all group across 1 directory with 6 updates (#480) [#480](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F480) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F5b33e3f1589d338ba18e5b3035794a7c8614a0fb))\n- **deps**: Bump kubectl from v1.32.1 to v1.32.2 in \u002Fcharts\u002Faikit [#484](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F484) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fa9561fc5a085e4a38235beb3c7302babba24e7a6))\n- **deps**: Bump kubectl from v1.32.1 to v1.32.2 in \u002Fcharts\u002Faikit (#484) [#484](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F484) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Faf0968cabcab0ee68c275ceb5db56f3bd466961c))\n- **deps**: Bump golang from 1.23-bookworm to 1.24-bookworm [#485](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002F","2025-02-22T06:24:58",{"id":249,"version":250,"summary_zh":251,"released_at":252},272084,"v0.16.0","## Notable Changes\r\n- ✨ Update to LocalAI [v2.24.1](https:\u002F\u002Fgithub.com\u002Fmudler\u002FLocalAI\u002Freleases\u002Ftag\u002Fv2.24.1).\r\n- 🖖 Apple Silicon containers uses [llama.cpp](https:\u002F\u002Fgithub.com\u002Fggerganov\u002Fllama.cpp) `kompute` backend instead of `vulkan`.\r\n- 🦙 Added Llama 3.3 70B and QwQ 32B to [pre-made models](https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fdocs\u002Fpremade-models).\r\n\r\n## Features\r\n- add qwq 32b preview (#443) [#443](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F443) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F98379ef3ba033a3586a79eaccb88bce8e0337950))\r\n- llama 3.3 70b (#448) [#448](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F448) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F8ad7653d008340dc2ee08a76cc915c2452f38189))\r\n- use kompute backend for apple silicon (#444) [#444](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F444) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F5b2a236643ea17eea79071d86b21dbe01d08c577))\r\n\r\n## Continuous Integration\r\n- allow storage endpoint (#440) [#440](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F440) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F74d4e2ff20c9807bc8b86a75bc99132f3bbe99d2))\r\n- fix name parsing (#449) [#449](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F449) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fbb0b1f40160d217071a1f4625a1df16c4ca3cbfc))\r\n\r\n## Chores\r\n- Prepare v0.15.0 release (#439) [#439](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F439) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fb26b7587ef9d180925eb8f873d273e75e4cc18ee))\r\n- **deps**: Bump github.com\u002Fmoby\u002Fbuildkit from 0.17.2 to 0.18.0 (#441) [#441](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F441) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F440954e98e4713f1b7d9a3bd15cd5b6a107fd221))\r\n- **deps**: Bump the actions-all group with 2 updates (#442) [#442](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F442) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F47a7c27688b4beb933c88f92fe8e6e28ae0b5e5e))\r\n- **deps**: Bump golang from `3f3b9da` to `1479eb3` (#445) [#445](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F445) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fc8573c892ff289708ba9f7b4d1fbd1702ad74b1b))\r\n- **deps**: Bump github.com\u002Fmoby\u002Fbuildkit from 0.18.0 to 0.18.1 (#446) [#446](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F446) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F18114ba44f99b57f8cc3dfa365d361179a71f9ea))\r\n- **deps**: Bump github\u002Fcodeql-action from 3.27.5 to 3.27.6 in the actions-all group (#447) [#447](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F447) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fd07f20c0f5004e7e92f4f24e5f8f1655dc3eb573))\r\n- **deps**: Bump golang.org\u002Fx\u002Fsync from 0.9.0 to 0.10.0 (#452) [#452](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F452) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F093662a47eb76b38630eece7ca9f4a71ac53f051))\r\n- **deps**: Bump golang from `1479eb3` to `ef30001` (#454) [#454](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F454) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Ff3b0a99da2b149d891badec246e934dfec984763))\r\n- **deps**: Bump the actions-all group with 2 updates (#453) [#453](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F453) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F06815375d5c23c898433614b31d02417c4740e48))\r\n- **deps**: Bump google.golang.org\u002Fgrpc from 1.68.0 to 1.68.1 (#451) [#451](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F451) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fa1fb47cd0abcbac4e42cd2b383b7d2eff4a8dac5))\r\n- bump localai v2.24.1 (#450) [#450](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F450) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F8f7a03672c6355671d0027b7e1d47425016c0847))\r\n- Prepare v0.16.0 release (#455) [#455](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F455) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fb2833a299d75461e6c11b5c30b60e0e4d11d257c))","2024-12-10T07:09:43",{"id":254,"version":255,"summary_zh":256,"released_at":257},272085,"v0.15.0","## Notable Changes\r\n- 🍎 [Apple Silicon GPU acceleration](https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fdocs\u002Fgpu#apple-silicon-experimental) through [Podman Desktop for macOS](https:\u002F\u002Fpodman-desktop.io\u002Fdocs\u002Fpodman\u002Fgpu)!\r\n\r\n\t- Get started quickly after setting up [Podman Desktop with GPUs (macOS)](https:\u002F\u002Fpodman-desktop.io\u002Fdocs\u002Fpodman\u002Fgpu) with [pre-made models](https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fdocs\u002Fpremade-models#apple-silicon-experimental) as easy as\r\n  \r\n\t```shell\r\n\tpodman run -d --rm --device \u002Fdev\u002Fdri -p 8080:8080 ghcr.io\u002Fsozercan\u002Fapplesilicon\u002Fllama3.1:8b\r\n\t```\r\n\r\n\tthen visit http:\u002F\u002Flocalhost:8080\u002Fchat or use:\r\n\t\r\n\t```shell\r\n\tcurl http:\u002F\u002Flocalhost:8080\u002Fv1\u002Fchat\u002Fcompletions -H \"Content-Type: application\u002Fjson\" -d '{\r\n\t    \"model\": \"llama-3.1-8b-instruct\",\r\n\t    \"messages\": [{\"role\": \"user\", \"content\": \"tell me about quantum mechanics\"}]\r\n\t  }'\r\n\t```\r\n\r\n   - Get started [creating your own images](https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fdocs\u002Fcreate-images) as easy as \r\n\r\n\t```shell\r\n\tdocker buildx build -t registry\u002Frepo\u002Fimage:tag --push \\\r\n\t    --build-arg=\"model=huggingface:\u002F\u002FMaziyarPanahi\u002FLlama-3.2-1B-Instruct-GGUF\u002FLlama-3.2-1B-Instruct.Q4_K_M.gguf\" \\\r\n\t    --build-arg=\"runtime=applesilicon\" \\\r\n\t    \"https:\u002F\u002Fraw.githubusercontent.com\u002Fsozercan\u002Faikit\u002Fmain\u002Fmodels\u002Faikitfile.yaml\"\r\n\t```\r\n\r\n- 🍏 [Pre-made models](https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fdocs\u002Fpremade-models#apple-silicon-experimental) for Llama 3.2 and 3.1, Phi 3.5, and Gemma 2 with Apple Silicon support\r\n- ✨ Update to LocalAI [v2.23.0](https:\u002F\u002Fgithub.com\u002Fmudler\u002FLocalAI\u002Freleases\u002Ftag\u002Fv2.23.0)\r\n- 🦥 Update Unsloth to [Sept 2024 release](https:\u002F\u002Fgithub.com\u002Funslothai\u002Funsloth\u002Freleases\u002Ftag\u002FSeptember-2024)\r\n\r\n## Features\r\n- apple silicon support (#433) [#433](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F433) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F95c9d54986fb4dc24488cfdaddabfa93bb4c5657))\r\n\r\n## Bug Fixes\r\n- set config file env (#401) [#401](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F401) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fb4d52b5b573ac15e14fe840f2ec19df65fbe4682))\r\n- fix llama 3.1 70b (#402) [#402](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F402) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F1b0a7f12cd1759350ce48b7318e68dda6c262042))\r\n- bump grpcio to 1.68.0 (#437) [#437](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F437) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F3a5f2344064d41a8d4e6e1f08e7fe2151b32b1d5))\r\n\r\n## Documentation\r\n- update apple silicon docs (#436) [#436](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F436) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fa23de47d5965305702878f0488aa53232ae115b5))\r\n\r\n## Continuous Integration\r\n- allow hf.co in update models (#400) [#400](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F400) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fa1c0cb1f97ce94a7dc52398720c9bf481f5a582c))\r\n- bump precommit (#421) [#421](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F421) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F136a0e93bb666e9924221ab398e55a03fe8a0643))\r\n- fix podman test (#435) [#435](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F435) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fbc55b5326eb893bd73421b5a61a72ad8ea8e591e))\r\n\r\n## Chores\r\n- Prepare v0.14.0 release (#399) [#399](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F399) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Ffdeb39fe55585b6aff3feef1e143d97ef076943d))\r\n- **deps**: Bump golang from `1a5326b` to `dba79eb` (#404) [#404](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F404) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F25f25f49e00beffd2470d0fb229913e79a9faa7a))\r\n- **deps**: Bump the actions-all group with 3 updates (#405) [#405](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F405) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F4e4ac66e45afed7a9b283efbb5a600bbb7868b78))\r\n- bump unsloth to sept-2024 (#403) [#403](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F403) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fbff6b84141127e331414c1c090b40ae622625c62))\r\n- **deps**: Bump google.golang.org\u002Fgrpc from 1.67.0 to 1.67.1 (#406) [#406](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F406) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fc542b493468dbc79956f065e2eda6818c0f8ca47))\r\n- **deps**: Bump the actions-all group with 6 updates (#407) [#407](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F407) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Ff73fd9d96707d3475b5690f2e76f706a14777c19))\r\n- **deps**: Bump golang from `dba79eb` to `18d2f94` (#408) [#408](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F408) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F8fddc75b896f17bb0a625038808850082998d65d))\r\n- **deps**: Bump golang from `18d2f94` to `2341ddf` (#410) [#410](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F410) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F73f6b2a941cd0d19e59d3ea17c1d9e228701bf05))\r\n- **deps**: Bump the actions-all group across 1 directory with 4 updates (#411) [#411](https:\u002F\u002Fgith","2024-11-27T07:39:42",{"id":259,"version":260,"summary_zh":261,"released_at":262},272086,"v0.14.0","## Notable Changes\r\n- 🦙 [Llama 3.2 1B and 3B](https:\u002F\u002Fai.meta.com\u002Fblog\u002Fllama-3-2-connect-2024-vision-edge-mobile-devices\u002F) are available in [pre-made images](https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fdocs\u002Fpremade-models)!\r\n\r\n## Features\r\n- add llama 3.2 (#395) [#395](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F395) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F72bcc36759b5c62c93e4c605e24f1b50d9cc5158))\r\n\r\n## Documentation\r\n- update flux emoji (#377) [#377](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F377) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F1b095668f2ee33544c18ad411de18c3d3e0f02ae))\r\n- add search (#386) [#386](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F386) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F7c5eb1efb4d92710abb179dafdd23bd8de84976f))\r\n\r\n## Continuous Integration\r\n- fix flux release (#378) [#378](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F378) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F2cb634856cd3fe62b6f3aa25c268c806e95f3856))\r\n- fix flux release (#379) [#379](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F379) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F7143e9529ff8bacf37b799bddeb3c9e3c875edf6))\r\n- update allowlist (#380) [#380](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F380) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F7ad5cae9d75d67dafc97bdb35eb0cf4f724b72f4))\r\n- update allowlist (#381) [#381](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F381) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F6ac149fe627c40a6aaba9e5e739f9629a1d1de3c))\r\n\r\n## Chores\r\n- Prepare v0.13.0 release (#376) [#376](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F376) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F914231079550f11e382dac80aac2277824a4ed83))\r\n- **deps**: Bump golang from `31dc846` to `1a5326b` (#384) [#384](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F384) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fc6245e818c9521a2f50c671a8e9cbc5ec677e661))\r\n- **deps**: Bump peter-evans\u002Fcreate-pull-request from 6.1.0 to 7.0.1 in the actions-all group (#383) [#383](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F383) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F8876e0bc07c67f95bd57ff1d7c1ec3c10016f110))\r\n- **deps**: Bump the npm_and_yarn group across 1 directory with 4 updates (#385) [#385](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F385) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fb4baa540b8804708b08dd513dcd9bde04123b86b))\r\n- **deps**: Bump kubectl from v1.31.0 to v1.31.1 in \u002Fcharts\u002Faikit (#388) [#388](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F388) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F78050d6db71fbad1503852bbdce5283d3809a5fb))\r\n- **deps**: Bump the actions-all group with 3 updates (#391) [#391](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F391) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F353f94ba43589fbfbb38e19055592ffc89db131f))\r\n- **deps**: Bump google.golang.org\u002Fgrpc from 1.66.0 to 1.67.0 (#393) [#393](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F393) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F6e9a9a985d3fe8962960ac6031acccc57cb64a02))\r\n- **deps**: Bump the actions-all group with 3 updates (#392) [#392](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F392) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F0cddd61d7d64b39ca8f5056f9985a419d6bbe5ad))\r\n- **deps**: Bump github.com\u002Fmoby\u002Fbuildkit from 0.15.2 to 0.16.0 (#389) [#389](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F389) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F8ed49c6e5a99c8c56079c72e74041fa18824e630))\r\n- remove deprecated exllama (#396) [#396](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F396) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F1fce06c2d6d5aa1f89e2fda00c093b9bc2bcf5fa))\r\n- update update-models script (#397) [#397](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F397) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F00874f1057ab7e462350d731aa3cd0e3b6664535))\r\n- Prepare v0.14.0 release (#398) [#398](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F398) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F9fa9ac1219c7832f555a680376ac1a7ec51b2216))","2024-09-27T03:50:54",{"id":264,"version":265,"summary_zh":266,"released_at":267},272087,"v0.13.0","## Notable Changes\r\n- 🤗 Added huggingface [`diffusers`](https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fdocs\u002Fdiffusion#diffusers) backend\r\n- 📸 [Flux-1 Dev](https:\u002F\u002Fblackforestlabs.ai\u002F#get-flux) is available in [pre-made images](https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fdocs\u002Fpremade-models)!\r\n- 🅿️ [Phi 3.5 Mini Instruct](https:\u002F\u002Fhuggingface.co\u002Fmicrosoft\u002FPhi-3.5-mini-instruct) is available in [pre-made images](https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fdocs\u002Fpremade-models)!\r\n\r\n## Features\r\n- add gemma2 2b (#355) [#355](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F355) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F5cfeed52d9e3b97b7f1ee8a5f05687e984118e6b))\r\n- diffusers backend (#373) [#373](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F373) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fb2d6722838a8e28432f8d1107dbe3a43b12bffa8))\r\n- add phi3.5 model (#374) [#374](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F374) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fc18d427f5249ec371e647d3bcfacc6988028b876))\r\n\r\n## Documentation\r\n- update for llama3.1 (#349) [#349](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F349) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F76ff007974a22b7cb9c7902ac7dec6bf8132edd8))\r\n\r\n## Continuous Integration\r\n- add ubuntu to allow list (#348) [#348](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F348) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fff77914fcc331494d616f060f454932ffadba162))\r\n- fix patch models (#358) [#358](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F358) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fc7dc5fd1d548e567e4cd61e6cde0d26cd9de1375))\r\n\r\n## Chores\r\n- Prepare v0.12.0 release (#347) [#347](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F347) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F3251792888da0e85daddb5d9497053e6304689c6))\r\n- **deps**: Bump github.com\u002Fmoby\u002Fbuildkit from 0.15.0 to 0.15.1 (#351) [#351](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F351) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F432977395448535b8c0fbb2fb883061dc9b4b014))\r\n- **deps**: Bump golang from `6c27802` to `af9b40f` (#352) [#352](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F352) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F39e98ae0ed887754cb29b26de331140e03a8b5c0))\r\n- **deps**: Bump the actions-all group with 6 updates (#350) [#350](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F350) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F6cbc2933eb2a3ba3ff0207d58cdc0c322249f8f4))\r\n- bump localai v2.19.4 (#343) [#343](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F343) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fba341a7595057203e3f3a494e1856e53c62a875c))\r\n- **deps**: Bump the actions-all group with 3 updates (#360) [#360](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F360) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F0fcd47ca2069ef4ab94c16ea7d8721a5124b0323))\r\n- **deps**: Bump golang.org\u002Fx\u002Fsync from 0.7.0 to 0.8.0 (#359) [#359](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F359) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fdaea225d63c5fa902cae896253c92dba86e7379b))\r\n- **deps**: Bump the actions-all group with 5 updates (#361) [#361](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F361) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fafafa6cf702553d617e916c88a309d002122d68e))\r\n- **deps**: Bump golang from `af9b40f` to `39b7e6e` (#362) [#362](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F362) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F35b05dcf93da1b49f283695042ce6a72dacec587))\r\n- **deps**: Bump golang from 1.22-bookworm to 1.23-bookworm (#364) [#364](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F364) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fd6e9d86b645ae847008b9816d0dbbd79dee34647))\r\n- **deps**: Bump the actions-all group with 2 updates (#366) [#366](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F366) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F4ab286a839fed9655e84fd4d8f7deb79b1a10c81))\r\n- **deps**: Bump github.com\u002Fmoby\u002Fbuildkit from 0.15.1 to 0.15.2 (#367) [#367](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F367) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F07c85229d2e94eed5f749070911aeb9e9b8353fa))\r\n- **deps**: Bump kubectl from v1.30.3 to v1.31.0 in \u002Fcharts\u002Faikit (#365) [#365](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F365) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fb5434a73bb2ffe1926982073553d3c1c5f476a00))\r\n- **deps**: Bump github\u002Fcodeql-action from 3.26.2 to 3.26.5 in the actions-all group (#368) [#368](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F368) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F4ddc24f959b5b16ebf290833706ab96e7876c4dd))\r\n- **deps**: Bump the npm_and_yarn group across 1 directory with 2 updates (#370) [#370](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F370) ([dependabot[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F5c1643e22dc99d9d28b5aa5c861e0911aa9b9340))\r\n- update localai to v2.20.1 (#369) [#369](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F369) ([Sertaç Özercan](https:\u002F\u002Fgithub.c","2024-09-07T20:35:44",{"id":269,"version":270,"summary_zh":271,"released_at":272},272088,"v0.12.2","## Notable Changes\r\n- 🔠 Gemma 2 2B is available in [pre-made images](https:\u002F\u002Fsozercan.github.io\u002Faikit\u002Fdocs\u002Fpremade-models)! \r\n\r\n## Chores\r\n- cherry-pick #355 (#356) [#356](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F356) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002Fda94e27dd454b6f3bfda39a556165b0970fcb473))\r\n- Prepare v0.12.2 release (#357) [#357](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F357) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F43058b2cab4df69c30d69a1d915f177ae1677bf7))","2024-08-03T23:18:44",{"id":274,"version":275,"summary_zh":276,"released_at":277},272089,"v0.12.1","## Notable Changes\r\n- 🆕 Update to LocalAI [v2.19.4](https:\u002F\u002Fgithub.com\u002Fmudler\u002FLocalAI\u002Freleases\u002Ftag\u002Fv2.19.4).\r\n\r\n## Chores\r\n- cherry pick #343 (#353) [#353](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F353) ([Sertaç Özercan](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F4a5b1a2ba9b0402e8c1b795d6bebcee5312b0a85))\r\n- Prepare v0.12.1 release (#354) [#354](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fpull\u002F354) ([github-actions[bot]](https:\u002F\u002Fgithub.com\u002Fsozercan\u002Faikit\u002Fcommit\u002F69284c690c497fb184fdf7cc21ec8770d4f07c31))","2024-08-03T20:26:36"]