base on Effortlessly run LLM backends, APIs, frontends, and services with one command. ![Harbor project logo](https://github.com/av/harbor/raw/main/docs/harbor-2.png) [![GitHub Tag](https://img.shields.io/github/v/tag/av/harbor)](https://github.com/av/harbor/releases) [![NPM Version](https://img.shields.io/npm/v/%40avcodes%2Fharbor?labelColor=red&color=white)](https://www.npmjs.com/package/@avcodes/harbor) [![PyPI - Version](https://img.shields.io/pypi/v/llm-harbor?labelColor=blue)](https://pypi.org/project/llm-harbor/) ![GitHub repo size](https://img.shields.io/github/repo-size/av/harbor) ![GitHub repo file or directory count](https://img.shields.io/github/directory-file-count/av/harbor?type=file&extension=yml&label=compose%20files&color=orange) [![Visitors](https://api.visitorbadge.io/api/visitors?path=av%2Fharbor&countColor=%23263759&style=flat)](https://visitorbadge.io/status?path=av%2Fharbor) ![GitHub language count](https://img.shields.io/github/languages/count/av/harbor) [![Discord](https://img.shields.io/badge/Discord-Harbor-blue?logo=discord&logoColor=white)](https://discord.gg/8nDRphrhSF) ![Harbor Ko-fi](https://img.shields.io/badge/Ko--fi-white?style=social&logo=kofi) Setup your local LLM stack effortlessly. ```bash # Starts fully configured Open WebUI and Ollama harbor up # Now, Open WebUI can do Web RAG and TTS/STT harbor up searxng speaches ``` Harbor is a CLI and companion app that lets you spin up a complete local LLM stack—backends like Ollama, llama.cpp, or vLLM, frontends like Open WebUI, plus supporting services like SearXNG for web search, Speaches for voice chat, and ComfyUI for image generation—all pre-wired to work together with a single `harbor up` command. No manual setup: just pick the services you want and Harbor handles the Docker Compose orchestration, configuration, and cross-service connectivity so you can focus on actually using your models. ![Screenshot of Harbor CLI and App together](https://github.com/av/harbor/wiki/harbor-app-3.png) ## Documentation - [Installing Harbor](https://github.com/av/harbor/wiki/1.0.-Installing-Harbor)<br/> Guides to install Harbor CLI and App - [Harbor User Guide](https://github.com/av/harbor/wiki/1.-Harbor-User-Guide)<br/> High-level overview of working with Harbor - [Harbor App](https://github.com/av/harbor/wiki/1.1-Harbor-App)<br/> Overview and manual for the Harbor companion application - [Harbor Services](https://github.com/av/harbor/wiki/2.-Services)<br/> Catalog of services available in Harbor - [Harbor CLI Reference](https://github.com/av/harbor/wiki/3.-Harbor-CLI-Reference)<br/> Read more about Harbor CLI commands and options. Read about supported services and the ways to configure them. - [Join our Discord](https://discord.gg/8nDRphrhSF)<br/> Get help, share your experience, and contribute to the project. ## What can Harbor do? ![Diagram outlining Harbor's service structure](https://raw.githubusercontent.com/wiki/av/harbor/harbor-arch-diag.png) #### ✦ Local LLMs Run LLMs and related services locally, with no or minimal configuration, typically in a single command or click. ```bash # All backends are pre-connected to Open WebUI harbor up ollama harbor up llamacpp harbor up vllm # Set and remember args for llama.cpp harbor llamacpp args -ngl 32 ``` #### Cutting Edge Inference Harbor supports most of the major inference engines as well as a few of the lesser-known ones. ```bash # We sincerely hope you'll never try to run all of them at once harbor up vllm llamacpp tgi litellm tabbyapi aphrodite sglang ktransformers mistralrs airllm ``` #### Tool Use Enjoy the benefits of MCP ecosystem, extend it to your use-cases. ```bash # Manage MCPs with a convenient Web UI harbor up metamcp # Connect MCPs to Open WebUI harbor up metamcp mcpo ``` #### Generate Images Harbor includes ComfyUI + Flux + Open WebUI integration. ```bash # Use FLUX in Open WebUI in one command harbor up comfyui ``` #### Local Web RAG / Deep Research Harbor includes [SearXNG](./docs/2.3.1-Satellite&colon-SearXNG.md) that is pre-connected to a lot of services out of the box: [Perplexica](./docs/2.3.2-Satellite&colon-Perplexica.md), [ChatUI](./docs/2.1.4-Frontend&colon-ChatUI.md), [Morphic](./docs/2.3.34-Satellite-Morphic.md), [Local Deep Research](./docs/2.3.45-Satellite-Local-Deep-Research.md) and more. ```bash # SearXNG is pre-connected to Open WebUI harbor up searxng # And to many other services harbor up searxng chatui harbor up searxng morphic harbor up searxng perplexica harbor up searxng ldr ``` #### LLM Workflows Harbor includes multiple services for build LLM-based data and chat workflows: [Dify](./docs/2.3.3-Satellite&colon-Dify.md), [LitLytics](./docs/2.3.21-Satellite&colon-LitLytics.md), [n8n](./docs/2.3.23-Satellite&colon-n8n.md), [Open WebUI Pipelines](./docs/2.3.25-Satellite&colon-Open-WebUI-Pipelines.md), [FloWise](./docs/2.3.31-Satellite&colon-Flowise.md), [LangFlow](./docs/2.3.32-Satellite&colon-LangFlow.md) ```bash # Use Dify in Open WebUI harbor up dify ``` #### Talk to your LLM Setup voice chats with your LLM in a single command. Open WebUI + Speaches ```bash # Speaches includes OpenAI-compatible SST and TTS # and connected to Open WebUI out of the box harbor up speaches ``` #### Chat from the phone You can access Harbor services from your phone with a QR code. Easily get links for local, LAN or Docker access. ```bash # Print a QR code to open the service on your phone harbor qr # Print a link to open the service on your phone harbor url webui ``` #### Chat from anywhere Harbor includes a [built-in tunneling service](./docs/3.-Harbor-CLI-Reference.md#harbor-tunnel-service) to expose your Harbor to the internet. > [!WARN] > Be careful exposing your computer to the Internet, it's not safe. ```bash # Expose default UI to the internet harbor tunnel # Expose a specific service to the internet # ⚠️ Ensure to configure authentication for the service harbor tunnel vllm # Harbor comes with traefik built-in and pre-configured # for all included services harbor up traefik ``` #### LLM Scripting [Harbor Boost](./docs/5.2.-Harbor-Boost.md) allows you to [easily script workflows](./docs/5.2.1.-Harbor-Boost-Custom-Modules.md) and interactions with downstream LLMs. ```bash # Use Harbor Boost to script LLM workflows harbor up boost ``` #### Config Profiles Save and manage configuration profiles for different scenarios. For example - save [llama.cpp](./docs/2.2.2-Backend&colon-llama.cpp.md) args for different models and contexts and switch between them easily. ```bash # Save and use config profiles harbor profile save llama4 harbor profile use default # Import profiles from a URL harbor profile use https://example.com/path/to/harbor-profile.env ``` #### Command History Harbor keeps a [local-only history of recent commands](./docs/3.-Harbor-CLI-Reference.md#harbor-history). Look up and re-run easily, standalone from the system shell history. ```bash # Lookup recently used harbor commands harbor history ``` #### Eject Ready to move to your own setup? Harbor [will give you](./docs/3.-Harbor-CLI-Reference.md#harbor-eject) a docker-compose file replicating your setup. ```bash # Eject from Harbor into a standalone Docker Compose setup # Will export related services and variables into a standalone file. harbor eject searxng llamacpp > docker-compose.harbor.yml ``` --- ## Services ##### UIs [Open WebUI](https://github.com/av/harbor/wiki/2.1.1-Frontend:-Open-WebUI) ⦁︎ [ComfyUI](https://github.com/av/harbor/wiki/2.1.2-Frontend:-ComfyUI) ⦁︎ [LibreChat](https://github.com/av/harbor/wiki/2.1.3-Frontend:-LibreChat) ⦁︎ [HuggingFace ChatUI](https://github.com/av/harbor/wiki/2.1.4-Frontend:-ChatUI) ⦁︎ [Lobe Chat](https://github.com/av/harbor/wiki/2.1.5-Frontend:-Lobe-Chat) ⦁︎ [Hollama](https://github.com/av/harbor/wiki/2.1.6-Frontend:-hollama) ⦁︎ [parllama](https://github.com/av/harbor/wiki/2.1.7-Frontend:-parllama) ⦁︎ [BionicGPT](https://github.com/av/harbor/wiki/2.1.8-Frontend:-BionicGPT) ⦁︎ [AnythingLLM](https://github.com/av/harbor/wiki/2.1.9-Frontend:-AnythingLLM) ⦁︎ [Chat Nio](https://github.com/av/harbor/wiki/2.1.10-Frontend:-Chat-Nio) ⦁︎ [mikupad](https://github.com/av/harbor/wiki/2.1.11-Frontend:-Mikupad) ⦁︎ [oterm](https://github.com/av/harbor/wiki/2.1.12-Frontend-oterm) ⦁︎ [omnichain](https://github.com/av/harbor/wiki/2.3.16-Satellite:-omnichain) ⦁︎ [ol1](https://github.com/av/harbor/wiki/2.3.19-Satellite:-ol1) ##### Backends [Ollama](https://github.com/av/harbor/wiki/2.2.1-Backend:-Ollama) ⦁︎ [llama.cpp](https://github.com/av/harbor/wiki/2.2.2-Backend:-llama.cpp) ⦁︎ [vLLM](https://github.com/av/harbor/wiki/2.2.3-Backend:-vLLM) ⦁︎ [TabbyAPI](https://github.com/av/harbor/wiki/2.2.4-Backend:-TabbyAPI) ⦁︎ [Aphrodite Engine](https://github.com/av/harbor/wiki/2.2.5-Backend:-Aphrodite-Engine) ⦁︎ [mistral.rs](https://github.com/av/harbor/wiki/2.2.6-Backend:-mistral.rs) ⦁︎ [openedai-speech](https://github.com/av/harbor/wiki/2.2.7-Backend:-openedai-speech) ⦁︎ [Speaches](https://github.com/av/harbor/wiki/2.2.14-Backend:-Speaches) ⦁︎ [Parler](https://github.com/av/harbor/wiki/2.2.8-Backend:-Parler) ⦁︎ [text-generation-inference](https://github.com/av/harbor/wiki/2.2.9-Backend:-text-generation-inference) ⦁︎ [LMDeploy](https://github.com/av/harbor/wiki/2.2.10-Backend:-lmdeploy) ⦁︎ [AirLLM](https://github.com/av/harbor/wiki/2.2.11-Backend:-AirLLM) ⦁︎ [SGLang](https://github.com/av/harbor/wiki/2.2.12-Backend:-SGLang) ⦁︎ [KTransformers](https://github.com/av/harbor/wiki/2.2.13-Backend:-KTransformers) ⦁︎ [Nexa SDK](https://github.com/av/harbor/wiki/2.2.15-Backend:-Nexa-SDK) ⦁︎ [KoboldCpp](https://github.com/av/harbor/wiki/2.2.16-Backend:-KoboldCpp) ⦁︎ [Modular MAX](https://github.com/av/harbor/wiki/2.2.17-Backend-Modular-MAX) ##### Satellites [Harbor Bench](https://github.com/av/harbor/wiki/5.1.-Harbor-Bench) ⦁︎ [Harbor Boost](https://github.com/av/harbor/wiki/5.2.-Harbor-Boost) ⦁︎ [SearXNG](https://github.com/av/harbor/wiki/2.3.1-Satellite:-SearXNG) ⦁︎ [Perplexica](https://github.com/av/harbor/wiki/2.3.2-Satellite:-Perplexica) ⦁︎ [Dify](https://github.com/av/harbor/wiki/2.3.3-Satellite:-Dify) ⦁︎ [Plandex](https://github.com/av/harbor/wiki/2.3.4-Satellite:-Plandex) ⦁︎ [LiteLLM](https://github.com/av/harbor/wiki/2.3.5-Satellite:-LiteLLM) ⦁︎ [LangFuse](https://github.com/av/harbor/wiki/2.3.6-Satellite:-langfuse) ⦁︎ [Open Interpreter](https://github.com/av/harbor/wiki/2.3.7-Satellite:-Open-Interpreter) ⦁ ︎[cloudflared](https://github.com/av/harbor/wiki/2.3.8-Satellite:-cloudflared) ⦁︎ [cmdh](https://github.com/av/harbor/wiki/2.3.9-Satellite:-cmdh) ⦁︎ [fabric](https://github.com/av/harbor/wiki/2.3.10-Satellite:-fabric) ⦁︎ [txtai RAG](https://github.com/av/harbor/wiki/2.3.11-Satellite:-txtai-RAG) ⦁︎ [TextGrad](https://github.com/av/harbor/wiki/2.3.12-Satellite:-TextGrad) ⦁︎ [Aider](https://github.com/av/harbor/wiki/2.3.13-Satellite:-aider) ⦁︎ [aichat](https://github.com/av/harbor/wiki/2.3.14-Satellite:-aichat) ⦁︎ [autogpt](https://github.com/av/harbor/wiki/2.3.15-Satellite:-AutoGPT) ⦁︎ [lm-evaluation-harness](https://github.com/av/harbor/wiki/2.3.17-Satellite:-lm-evaluation-harness) ⦁︎ [JupyterLab](https://github.com/av/harbor/wiki/2.3.18-Satellite:-JupyterLab) ⦁︎ [ol1](https://github.com/av/harbor/wiki/2.3.19-Satellite:-ol1) ⦁︎ [OpenHands](https://github.com/av/harbor/wiki/2.3.20-Satellite:-OpenHands) ⦁︎ [LitLytics](https://github.com/av/harbor/wiki/2.3.21-Satellite:-LitLytics) ⦁︎ [Repopack](https://github.com/av/harbor/wiki/2.3.22-Satellite:-Repopack) ⦁︎ [n8n](https://github.com/av/harbor/wiki/2.3.23-Satellite:-n8n) ⦁︎ [Bolt.new](https://github.com/av/harbor/wiki/2.3.24-Satellite:-Bolt.new) ⦁︎ [Open WebUI Pipelines](https://github.com/av/harbor/wiki/2.3.25-Satellite:-Open-WebUI-Pipelines) ⦁︎ [Qdrant](https://github.com/av/harbor/wiki/2.3.26-Satellite:-Qdrant) ⦁︎ [K6](https://github.com/av/harbor/wiki/2.3.27-Satellite:-K6) ⦁︎ [Promptfoo](https://github.com/av/harbor/wiki/2.3.28-Satellite:-Promptfoo) ⦁︎ [Webtop](https://github.com/av/harbor/wiki/2.3.29-Satellite:-Webtop) ⦁︎ [OmniParser](https://github.com/av/harbor/wiki/2.3.30-Satellite:-OmniParser) ⦁︎ [Flowise](https://github.com/av/harbor/wiki/2.3.31-Satellite:-Flowise) ⦁︎ [Langflow](https://github.com/av/harbor/wiki/2.3.32-Satellite:-LangFlow) ⦁︎ [OptiLLM](https://github.com/av/harbor/wiki/2.3.33-Satellite:-OptiLLM) ⦁︎ [Morphic](https://github.com/av/harbor/wiki/2.3.34-Satellite-Morphic) ⦁︎ [SQL Chat](https://github.com/av/harbor/wiki/2.3.35-Satellite-SQL-Chat) ⦁︎ [gptme](https://github.com/av/harbor/wiki/2.3.36-Satellite-gptme) ⦁︎ [traefik](https://github.com/av/harbor/wiki/2.3.37-Satellite-traefik) ⦁︎ [Latent Scope](https://github.com/av/harbor/wiki/2.3.38-Satellite-Latent-Scope) ⦁︎ [RAGLite](https://github.com/av/harbor/wiki/2.3.39-Satellite-RAGLite) ⦁︎ [llama-swap](https://github.com/av/harbor/wiki/2.3.40-Satellite-llamaswap) ⦁︎ [LibreTranslate](https://github.com/av/harbor/wiki/2.3.41-Satellite-LibreTranslate) ⦁︎ [MetaMCP](https://github.com/av/harbor/wiki/2.3.42-Satellite-MetaMCP) ⦁︎ [mcpo](https://github.com/av/harbor/wiki/2.3.43-Satellite-mcpo) ⦁︎ [SuperGateway](https://github.com/av/harbor/wiki/2.3.44-Satellite-supergateway) ⦁︎ [Local Deep Research](https://github.com/av/harbor/wiki/2.3.45-Satellite-Local-Deep-Research) ⦁︎ [LocalAI](https://github.com/av/harbor/wiki/2.3.46-Satellite-LocalAI) ⦁︎ [AgentZero](https://github.com/av/harbor/wiki/2.3.47-Satellite-Agent-Zero) ⦁︎ [Airweave](https://github.com/av/harbor/wiki/2.3.48-Satellite-Airweave) ⦁︎ [Docling](https://github.com/av/harbor/wiki/2.3.49-Satellite-Docling) ⦁︎ [Browser Use](https://github.com/av/harbor/wiki/2.3.50-Satellite-Browser-Use) ⦁︎ [Unsloth](https://github.com/av/harbor/wiki/2.3.51-Satellite-Unsloth) ⦁︎ [Windmill](https://github.com/av/harbor/wiki/2.3.52-Satellite-Windmill) See [services documentation](https://github.com/av/harbor/wiki/2.-Services) for a brief overview of each. ## CLI Tour ```bash # Run Harbor with default services: # Open WebUI and Ollama harbor up # Run Harbor with additional services # Running SearXNG automatically enables Web RAG in Open WebUI harbor up searxng # Speaches includes OpenAI-compatible SST and TTS # and connected to Open WebUI out of the box harbor up speaches # Run additional/alternative LLM Inference backends # Open Webui is automatically connected to them. harbor up llamacpp tgi litellm vllm tabbyapi aphrodite sglang ktransformers # Run different Frontends harbor up librechat chatui bionicgpt hollama # Get a free quality boost with # built-in optimizing proxy harbor up boost # Use FLUX in Open WebUI in one command harbor up comfyui # Use custom models for supported backends harbor llamacpp model https://huggingface.co/user/repo/model.gguf # Access service CLIs without installing them # Caches are shared between services where possible harbor hf scan-cache harbor hf download google/gemma-2-2b-it harbor ollama list # Shortcut to HF Hub to find the models harbor hf find gguf gemma-2 # Use HFDownloader and official HF CLI to download models harbor hf dl -m google/gemma-2-2b-it -c 10 -s ./hf harbor hf download google/gemma-2-2b-it # Where possible, cache is shared between the services harbor tgi model google/gemma-2-2b-it harbor vllm model google/gemma-2-2b-it harbor aphrodite model google/gemma-2-2b-it harbor tabbyapi model google/gemma-2-2b-it-exl2 harbor mistralrs model google/gemma-2-2b-it harbor opint model google/gemma-2-2b-it harbor sglang model google/gemma-2-2b-it # Convenience tools for docker setup harbor logs llamacpp harbor exec llamacpp ./scripts/llama-bench --help harbor shell vllm # Tell your shell exactly what you think about it harbor opint harbor aider harbor aichat harbor cmdh # Use fabric to LLM-ify your linux pipes cat ./file.md | harbor fabric --pattern extract_extraordinary_claims | grep "LK99" # Open services from the CLI harbor open webui harbor open llamacpp # Print yourself a QR to quickly open the # service on your phone harbor qr # Feeling adventurous? Expose your Harbor # to the internet harbor tunnel # Config management harbor config list harbor config set webui.host.port 8080 # Create and manage config profiles harbor profile save l370b harbor profile use default # Import profile from a URL harbor profile use https://example.com/path/to/harbor-profile.env # Lookup recently used harbor commands harbor history # Eject from Harbor into a standalone Docker Compose setup # Will export related services and variables into a standalone file. harbor eject searxng llamacpp > docker-compose.harbor.yml # Run a built-in LLM benchmark with # your own tasks harbor bench run # Gimmick/Fun Area # Argument scrambling, below commands are all the same as above # Harbor doesn't care if it's "vllm model" or "model vllm", it'll # figure it out. harbor model vllm harbor vllm model harbor config get webui.name harbor get config webui_name harbor tabbyapi shell harbor shell tabbyapi # 50% gimmick, 50% useful # Ask harbor about itself harbor how to ping ollama container from the webui? ``` ## Harbor App Demo https://github.com/user-attachments/assets/a5cd2ef1-3208-400a-8866-7abd85808503 In the demo, Harbor App is used to launch a default stack with [Ollama](./2.2.1-Backend:-Ollama) and [Open WebUI](./2.1.1-Frontend:-Open-WebUI) services. Later, [SearXNG](./2.3.1-Satellite:-SearXNG) is also started, and WebUI can connect to it for the Web RAG right out of the box. After that, [Harbor Boost](./5.2.-Harbor-Boost) is also started and connected to the WebUI automatically to induce more creative outputs. As a final step, Harbor config is adjusted in the App for the [`klmbr`](./5.2.-Harbor-Boost#klmbr---boost-llm-creativity) module in the [Harbor Boost](./5.2.-Harbor-Boost), which makes the output unparsable for the LLM (yet still undetstandable for humans). ## Why? - If you're comfortable with Docker and Linux administration - you likely don't need Harbor to manage your local LLM environment. However, while growing it - you're also likely to eventually arrive to a similar solution. I know this for a fact, since that's exactly how Harbor came to be. - Harbor is not designed as a deployment solution, but rather as a helper for the local LLM development environment. It's a good starting point for experimenting with LLMs and related services. - Workflow/setup centralisation - you can be sure where to find a specific config or service, logs, data and configuration files. - Convenience factor - single CLI with a lot of services and features, accessible from anywhere on your host. ## Supporters ![@av's wife](https://ui-avatars.com/api/?size=32&name=KN&rounded=true&background=ffaaaa&color=ff4444) ![@burnth3heretic](https://ui-avatars.com/api/?size=32&name=BTH&rounded=true) ![@vood](https://ui-avatars.com/api/?size=32&name=VD&rounded=true) ![@anonymous](https://ui-avatars.com/api/?size=32&name=🥷&rounded=true&background=bada55) ", Assign "at most 3 tags" to the expected json: {"id":"12658","tags":[]} "only from the tags list I provide: [{"id":77,"name":"3d"},{"id":89,"name":"agent"},{"id":17,"name":"ai"},{"id":54,"name":"algorithm"},{"id":24,"name":"api"},{"id":44,"name":"authentication"},{"id":3,"name":"aws"},{"id":27,"name":"backend"},{"id":60,"name":"benchmark"},{"id":72,"name":"best-practices"},{"id":39,"name":"bitcoin"},{"id":37,"name":"blockchain"},{"id":1,"name":"blog"},{"id":45,"name":"bundler"},{"id":58,"name":"cache"},{"id":21,"name":"chat"},{"id":49,"name":"cicd"},{"id":4,"name":"cli"},{"id":64,"name":"cloud-native"},{"id":48,"name":"cms"},{"id":61,"name":"compiler"},{"id":68,"name":"containerization"},{"id":92,"name":"crm"},{"id":34,"name":"data"},{"id":47,"name":"database"},{"id":8,"name":"declarative-gui "},{"id":9,"name":"deploy-tool"},{"id":53,"name":"desktop-app"},{"id":6,"name":"dev-exp-lib"},{"id":59,"name":"dev-tool"},{"id":13,"name":"ecommerce"},{"id":26,"name":"editor"},{"id":66,"name":"emulator"},{"id":62,"name":"filesystem"},{"id":80,"name":"finance"},{"id":15,"name":"firmware"},{"id":73,"name":"for-fun"},{"id":2,"name":"framework"},{"id":11,"name":"frontend"},{"id":22,"name":"game"},{"id":81,"name":"game-engine "},{"id":23,"name":"graphql"},{"id":84,"name":"gui"},{"id":91,"name":"http"},{"id":5,"name":"http-client"},{"id":51,"name":"iac"},{"id":30,"name":"ide"},{"id":78,"name":"iot"},{"id":40,"name":"json"},{"id":83,"name":"julian"},{"id":38,"name":"k8s"},{"id":31,"name":"language"},{"id":10,"name":"learning-resource"},{"id":33,"name":"lib"},{"id":41,"name":"linter"},{"id":28,"name":"lms"},{"id":16,"name":"logging"},{"id":76,"name":"low-code"},{"id":90,"name":"message-queue"},{"id":42,"name":"mobile-app"},{"id":18,"name":"monitoring"},{"id":36,"name":"networking"},{"id":7,"name":"node-version"},{"id":55,"name":"nosql"},{"id":57,"name":"observability"},{"id":46,"name":"orm"},{"id":52,"name":"os"},{"id":14,"name":"parser"},{"id":74,"name":"react"},{"id":82,"name":"real-time"},{"id":56,"name":"robot"},{"id":65,"name":"runtime"},{"id":32,"name":"sdk"},{"id":71,"name":"search"},{"id":63,"name":"secrets"},{"id":25,"name":"security"},{"id":85,"name":"server"},{"id":86,"name":"serverless"},{"id":70,"name":"storage"},{"id":75,"name":"system-design"},{"id":79,"name":"terminal"},{"id":29,"name":"testing"},{"id":12,"name":"ui"},{"id":50,"name":"ux"},{"id":88,"name":"video"},{"id":20,"name":"web-app"},{"id":35,"name":"web-server"},{"id":43,"name":"webassembly"},{"id":69,"name":"workflow"},{"id":87,"name":"yaml"}]" returns me the "expected json"