Featured

AI infrastructure

New 2026

DeepSeek 4 Flash local inference engine for Metal

New 2026

A kernel library written in tilelang

New 2026

将运行在本地的 OpenCode 转换为 OpenAI 兼容 API ,以在任何 OpenAI 客户端中使用免费模型。

New 2026

The ultimate collection of APIs for building autonomous AI agents — 2,036 production-ready APIs across Agents, AI Models, and MCP Servers. Stop wasting weeks building infrastructure. Plug these in and ship your agent today.

New 2026

Garry's Opinionated OpenClaw/Hermes Agent Brain

New 2026

TriAttention — Efficient long reasoning with trigonometric KV cache compression. Enables OpenClaw local deployment on memory-constrained GPUs.

New 2026

Lucebox optimization hub: hand-tuned LLM inference, built for specific consumer hardware.

New 2026

The open-source memory operating system for AI agents. Persistent memory, semantic search, loop detection, agent messaging, crash recovery, and real-time observability.

New 2026

AI API identity gateway — reverse proxy that normalizes device fingerprints and telemetry for privacy-preserving API proxying

New 2026

Awesome list for AI agent harness engineering: tools, patterns, evals, memory, MCP, permissions, observability, and orchestration.

New 2026

TurboQuant: Near-optimal KV cache quantization for LLM inference (3-bit keys, 2-bit values) with Triton kernels + vLLM integration

New 2026

Zig INferenCe Engine — Local LLM inference on AMD GPUs and Apple Silicon

New 2026

The open-source AI platform for enterprises that can't send data to the cloud. OpenAI-compatible API, full management dashboard, zero data egress.

New 2026

Curated list of the best truly open-source AI projects, models, tools, and infrastructure.

New 2026

OpsKat — Infrastructure Ops, Reimagined with AI

New 2026

⚡ Native MLX Swift LLM inference server for Apple Silicon. OpenAI-compatible API, SSD streaming for 100B+ MoE models, TurboQuant KV cache compression, MACOS + iOS iPhone app.

New 2026

Self-hosted AI Agent Memory + Code Intelligence Platform — one MCP endpoint for persistent memory, AST-aware code search, shared knowledge, and quality enforcement across all your AI coding agents.

New 2026

Privacy-first LLM proxy and AI gateway — load balancing, multi-provider routing, API key management, usage tracking, rate limiting. Self-hosted. Zero knowledge of your prompts.

New 2026

CORAL is a robust, lightweight infrastructure for multi-agent autonomous self-evolution, built for autoresearch.