diff --git a/.gitignore b/.gitignore index a216a5c..c818df7 100644 --- a/.gitignore +++ b/.gitignore @@ -1,3 +1,5 @@ +supervisor.conf +supervisor .venv venv config.json @@ -5,4 +7,4 @@ bins/mcphost mcphost_openai_api.log.* commons/settings.py .idea -__pycache__ \ No newline at end of file +__pycache__ diff --git a/commons/logging_utils.py b/commons/logging_utils.py index b9dd5fd..b36f51a 100644 --- a/commons/logging_utils.py +++ b/commons/logging_utils.py @@ -40,4 +40,4 @@ def setup_logger(): diagnose=True ) - return logger \ No newline at end of file + return logger diff --git a/commons/mcp_manager.py b/commons/mcp_manager.py index 9a994fa..84eefda 100644 --- a/commons/mcp_manager.py +++ b/commons/mcp_manager.py @@ -181,4 +181,4 @@ class MCPHostManager: break logger.debug("Collected response length: {} characters", len(response)) - return response \ No newline at end of file + return response diff --git a/logs.sh b/logs.sh new file mode 100644 index 0000000..cbf96a6 --- /dev/null +++ b/logs.sh @@ -0,0 +1,3 @@ +#!/bin/bash + +tail -f supervisor/logs/llm-api-mcphost.log diff --git a/serve_mcphost_openai_compatible.py b/serve_mcphost_openai_compatible.py index 394e446..1f22408 100644 --- a/serve_mcphost_openai_compatible.py +++ b/serve_mcphost_openai_compatible.py @@ -124,4 +124,4 @@ async def health_check(): if __name__ == "__main__": import uvicorn - uvicorn.run(app, host="0.0.0.0", port=8000) \ No newline at end of file + uvicorn.run(app, host=settings.host, port=settings.port)