From c75fa97cb7a336f7a50726395a4e5e2d38241f5b Mon Sep 17 00:00:00 2001 From: Pavan Kumar <66913595+ascender1729@users.noreply.github.com> Date: Thu, 17 Apr 2025 17:21:24 +0000 Subject: [PATCH] fix: resolve LiteLLM initialization issues in math and calc servers - Updated initialization logic in `math_server.py` and `calc_server.py` - Addressed missing or invalid response output due to improper model handling --- examples/mcp_example/calc_server.py | 2 +- examples/mcp_example/math_server.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/examples/mcp_example/calc_server.py b/examples/mcp_example/calc_server.py index 25161ec3..07d7391b 100644 --- a/examples/mcp_example/calc_server.py +++ b/examples/mcp_example/calc_server.py @@ -29,5 +29,5 @@ def percentage(value: float, percent: float) -> float: if __name__ == "__main__": print("Starting Calculation Server on port 6275...") - llm = LiteLLM() + llm = LiteLLM(system_prompt="You are a financial calculation expert.") mcp.run(transport="sse", host="0.0.0.0", port=6275) diff --git a/examples/mcp_example/math_server.py b/examples/mcp_example/math_server.py index 9fcfcafd..e2b25156 100644 --- a/examples/mcp_example/math_server.py +++ b/examples/mcp_example/math_server.py @@ -66,5 +66,5 @@ def divide(a: float, b: float) -> float: if __name__ == "__main__": print("Starting Math Server on port 6274...") - llm = LiteLLM(model_name="gpt-4o-mini") + llm = LiteLLM() # This will use the default model mcp.run(transport="sse", host="0.0.0.0", port=6274) \ No newline at end of file