Optimizing inference proxy for LLMs
llm
large-language-models
openai
prompt-engineering
agent
agents
proxy-server
genai
llm-inference
agentic-ai
optimization
agentic-framework
agentic-workflow
api-gateway
chain-of-thought
llmapi
mixture-of-experts
moa
monte-carlo-tree-search
openai-api
Updated 2025-05-28 09:39:38 +03:00
Caddy server 2.0.0 / 1.0.5 on Docker Scratch, all in 18MB / 35MB
Updated 2022-03-12 19:25:16 +03:00