diff --git a/app.py b/app.py index 6ddf8fee..3a696590 100644 --- a/app.py +++ b/app.py @@ -8,26 +8,20 @@ import warnings from swarms.modelui.modules.block_requests import OpenMonkeyPatch, RequestBlocker from swarms.modelui.modules.logging_colors import logger -from vllm import LLM +from vllm import LLM -os.environ["GRADIO_ANALYTICS_ENABLED"] = "False" -os.environ["BITSANDBYTES_NOWELCOME"] = "1" -warnings.filterwarnings( - "ignore", category=UserWarning, message="TypedStorage is deprecated" -) -warnings.filterwarnings( - "ignore", category=UserWarning, message="Using the update method is deprecated" -) -warnings.filterwarnings( - "ignore", category=UserWarning, message='Field "model_name" has conflict' -) +os.environ['GRADIO_ANALYTICS_ENABLED'] = 'False' +os.environ['BITSANDBYTES_NOWELCOME'] = '1' +warnings.filterwarnings('ignore', category=UserWarning, message='TypedStorage is deprecated') +warnings.filterwarnings('ignore', category=UserWarning, message='Using the update method is deprecated') +warnings.filterwarnings('ignore', category=UserWarning, message='Field "model_name" has conflict') with RequestBlocker(): import gradio as gr import matplotlib -matplotlib.use("Agg") # This fixes LaTeX rendering on some systems +matplotlib.use('Agg') # This fixes LaTeX rendering on some systems import swarms.modelui.modules.extensions as extensions_module from swarms.modelui.modules import ( @@ -42,7 +36,7 @@ from swarms.modelui.modules import ( ui_notebook, ui_parameters, ui_session, - utils, + utils ) from swarms.modelui.modules.extensions import apply_extensions from swarms.modelui.modules.LoRA import add_lora_to_model @@ -50,7 +44,7 @@ from swarms.modelui.modules.models import load_model from swarms.modelui.modules.models_settings import ( get_fallback_settings, get_model_metadata, - update_model_parameters, + update_model_parameters ) from swarms.modelui.modules.utils import gradio @@ -68,8 +62,6 @@ import time from langchain.llms import VLLM tool_server_flag = False - - def start_tool_server(): # server = Thread(target=run_tool_server) server = Process(target=run_tool_server) @@ -78,9 +70,12 @@ def start_tool_server(): tool_server_flag = True -available_models = ["ChatGPT", "GPT-3.5"] DEFAULTMODEL = "ChatGPT" # "GPT-3.5" +# Read the model/ directory and get the list of models +model_dir = Path("./models/") +available_models = ["ChatGPT", "GPT-3.5"] + [f.name for f in model_dir.iterdir() if f.is_dir()] + tools_mappings = { "klarna": "https://www.klarna.com/", "weather": "http://127.0.0.1:8079/tools/weather/", @@ -132,23 +127,21 @@ chat_history = "" MAX_SLEEP_TIME = 40 - def download_model(model_url: str, memory_utilization: int): # Extract model name from the URL - model_name = model_url.split("/")[-1] + model_name = model_url.split('/')[-1] # Download the model using VLLM vllm_model = VLLM( model=model_url, trust_remote_code=True, gpu_memory_utilization=memory_utilization, - download_dir="models/", + download_dir=model_dir ) # Add the downloaded model to the available_models list available_models.append((model_name, vllm_model)) # Update the dropdown choices with the new available_models list model_chosen.update(choices=available_models) - def load_tools(): global valid_tools_info global all_tools_list @@ -159,24 +152,23 @@ def load_tools(): all_tools_list = sorted(list(valid_tools_info.keys())) return gr.update(choices=all_tools_list) +def set_environ(OPENAI_API_KEY: str = "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", + WOLFRAMALPH_APP_ID: str = "", + WEATHER_API_KEYS: str = "", + BING_SUBSCRIPT_KEY: str = "", + ALPHA_VANTAGE_KEY: str = "", + BING_MAP_KEY: str = "", + BAIDU_TRANSLATE_KEY: str = "", + RAPIDAPI_KEY: str = "", + SERPER_API_KEY: str = "", + GPLACES_API_KEY: str = "", + SCENEX_API_KEY: str = "", + STEAMSHIP_API_KEY: str = "", + HUGGINGFACE_API_KEY: str = "", + AMADEUS_ID: str = "", + AMADEUS_KEY: str = "", + ): -def set_environ( - OPENAI_API_KEY: str = "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", - WOLFRAMALPH_APP_ID: str = "", - WEATHER_API_KEYS: str = "", - BING_SUBSCRIPT_KEY: str = "", - ALPHA_VANTAGE_KEY: str = "", - BING_MAP_KEY: str = "", - BAIDU_TRANSLATE_KEY: str = "", - RAPIDAPI_KEY: str = "", - SERPER_API_KEY: str = "", - GPLACES_API_KEY: str = "", - SCENEX_API_KEY: str = "", - STEAMSHIP_API_KEY: str = "", - HUGGINGFACE_API_KEY: str = "", - AMADEUS_ID: str = "", - AMADEUS_KEY: str = "", -): os.environ["OPENAI_API_KEY"] = OPENAI_API_KEY os.environ["WOLFRAMALPH_APP_ID"] = WOLFRAMALPH_APP_ID os.environ["WEATHER_API_KEYS"] = WEATHER_API_KEYS @@ -197,53 +189,31 @@ def set_environ( time.sleep(MAX_SLEEP_TIME) return gr.update(value="OK!") - def show_avatar_imgs(tools_chosen): if len(tools_chosen) == 0: tools_chosen = list(valid_tools_info.keys()) img_template = ' avatar {} ' - imgs = [ - valid_tools_info[tool]["avatar"] - for tool in tools_chosen - if valid_tools_info[tool]["avatar"] != None - ] - imgs = " ".join( - [img_template.format(img, img, tool) for img, tool in zip(imgs, tools_chosen)] - ) - return [ - gr.update(value='' + imgs + "", visible=True), - gr.update(visible=True), - ] - + imgs = [valid_tools_info[tool]['avatar'] for tool in tools_chosen if valid_tools_info[tool]['avatar'] != None] + imgs = ' '.join([img_template.format(img, img, tool) for img, tool in zip(imgs, tools_chosen)]) + return [gr.update(value='' + imgs + '', visible=True), gr.update(visible=True)] def answer_by_tools(question, tools_chosen, model_chosen): global return_msg - return_msg += [(question, None), (None, "...")] + return_msg += [(question, None), (None, '...')] yield [gr.update(visible=True, value=return_msg), gr.update(), gr.update()] - OPENAI_API_KEY = os.environ.get("OPENAI_API_KEY", "") + OPENAI_API_KEY = os.environ.get('OPENAI_API_KEY', '') - if ( - len(tools_chosen) == 0 - ): # if there is no tools chosen, we use all todo (TODO: What if the pool is too large.) + if len(tools_chosen) == 0: # if there is no tools chosen, we use all todo (TODO: What if the pool is too large.) tools_chosen = list(valid_tools_info.keys()) if len(tools_chosen) == 1: - answerer = STQuestionAnswerer( - OPENAI_API_KEY.strip(), stream_output=True, llm=model_chosen - ) - agent_executor = answerer.load_tools( - tools_chosen[0], - valid_tools_info[tools_chosen[0]], - prompt_type="react-with-tool-description", - return_intermediate_steps=True, - ) + answerer = STQuestionAnswerer(OPENAI_API_KEY.strip(), stream_output=True, llm=model_chosen) + agent_executor = answerer.load_tools(tools_chosen[0], valid_tools_info[tools_chosen[0]], + prompt_type="react-with-tool-description", return_intermediate_steps=True) else: - answerer = MTQuestionAnswerer( - OPENAI_API_KEY.strip(), - load_valid_tools({k: tools_mappings[k] for k in tools_chosen}), - stream_output=True, - llm=model_chosen, - ) + answerer = MTQuestionAnswerer(OPENAI_API_KEY.strip(), + load_valid_tools({k: tools_mappings[k] for k in tools_chosen}), + stream_output=True, llm=model_chosen) agent_executor = answerer.build_runner() @@ -251,15 +221,12 @@ def answer_by_tools(question, tools_chosen, model_chosen): chat_history += "Question: " + question + "\n" question = chat_history for inter in agent_executor(question): - if isinstance(inter, AgentFinish): - continue + if isinstance(inter, AgentFinish): continue result_str = [] return_msg.pop() if isinstance(inter, dict): - result_str.append( - "Answer: {}".format(inter["output"]) - ) - chat_history += "Answer:" + inter["output"] + "\n" + result_str.append("Answer: {}".format(inter['output'])) + chat_history += "Answer:" + inter['output'] + "\n" result_str.append("...") else: try: @@ -267,40 +234,23 @@ def answer_by_tools(question, tools_chosen, model_chosen): except: print(inter[0]) not_observation = inter[0] - if not not_observation.startswith("Thought:"): + if not not_observation.startswith('Thought:'): not_observation = "Thought: " + not_observation chat_history += not_observation - not_observation = not_observation.replace( - "Thought:", "Thought: " - ) - not_observation = not_observation.replace( - "Action:", "Action: " - ) - not_observation = not_observation.replace( - "Action Input:", "Action Input: " - ) + not_observation = not_observation.replace('Thought:', 'Thought: ') + not_observation = not_observation.replace('Action:', 'Action: ') + not_observation = not_observation.replace('Action Input:', 'Action Input: ') result_str.append("{}".format(not_observation)) - result_str.append( - "Action output:\n{}".format(inter[1]) - ) + result_str.append("Action output:\n{}".format(inter[1])) chat_history += "\nAction output:" + inter[1] + "\n" result_str.append("...") return_msg += [(None, result) for result in result_str] yield [gr.update(visible=True, value=return_msg), gr.update(), gr.update()] return_msg.pop() if return_msg[-1][1].startswith("Answer: "): - return_msg[-1] = ( - return_msg[-1][0], - return_msg[-1][1].replace( - "Answer: ", - "Final Answer: ", - ), - ) - yield [ - gr.update(visible=True, value=return_msg), - gr.update(visible=True), - gr.update(visible=False), - ] + return_msg[-1] = (return_msg[-1][0], return_msg[-1][1].replace("Answer: ", + "Final Answer: ")) + yield [gr.update(visible=True, value=return_msg), gr.update(visible=True), gr.update(visible=False)] def retrieve(tools_search): @@ -308,13 +258,14 @@ def retrieve(tools_search): return gr.update(choices=all_tools_list) else: url = "http://127.0.0.1:8079/retrieve" - param = {"query": tools_search} + param = { + "query": tools_search + } response = requests.post(url, json=param) result = response.json() retrieved_tools = result["tools"] return gr.update(choices=retrieved_tools) - def clear_retrieve(): return [gr.update(value=""), gr.update(choices=all_tools_list)] @@ -326,14 +277,13 @@ def clear_history(): chat_history = "" yield gr.update(visible=True, value=return_msg) - -title = "Swarm Models" +title = 'Swarm Models' # css/js strings css = ui.css js = ui.js -css += apply_extensions("css") -js += apply_extensions("js") +css += apply_extensions('css') +js += apply_extensions('js') # with gr.Blocks(css=css, analytics_enabled=False, title=title, theme=ui.theme) as demo: with gr.Blocks() as demo: @@ -341,93 +291,39 @@ with gr.Blocks() as demo: with gr.Column(scale=14): gr.Markdown("") with gr.Column(scale=1): - gr.Image( - show_label=False, - show_download_button=False, - value="images/swarmslogobanner.png", - ) + gr.Image(show_label=False, show_download_button=False, value="images/swarmslogobanner.png") with gr.Tab("Key setting"): - OPENAI_API_KEY = gr.Textbox( - label="OpenAI API KEY:", placeholder="sk-...", type="text" - ) - WOLFRAMALPH_APP_ID = gr.Textbox( - label="Wolframalpha app id:", - placeholder="Key to use wlframalpha", - type="text", - ) - WEATHER_API_KEYS = gr.Textbox( - label="Weather api key:", placeholder="Key to use weather api", type="text" - ) - BING_SUBSCRIPT_KEY = gr.Textbox( - label="Bing subscript key:", - placeholder="Key to use bing search", - type="text", - ) - ALPHA_VANTAGE_KEY = gr.Textbox( - label="Stock api key:", placeholder="Key to use stock api", type="text" - ) - BING_MAP_KEY = gr.Textbox( - label="Bing map key:", placeholder="Key to use bing map", type="text" - ) - BAIDU_TRANSLATE_KEY = gr.Textbox( - label="Baidu translation key:", - placeholder="Key to use baidu translation", - type="text", - ) - RAPIDAPI_KEY = gr.Textbox( - label="Rapidapi key:", - placeholder="Key to use zillow, airbnb and job search", - type="text", - ) - SERPER_API_KEY = gr.Textbox( - label="Serper key:", - placeholder="Key to use google serper and google scholar", - type="text", - ) - GPLACES_API_KEY = gr.Textbox( - label="Google places key:", - placeholder="Key to use google places", - type="text", - ) - SCENEX_API_KEY = gr.Textbox( - label="Scenex api key:", placeholder="Key to use sceneXplain", type="text" - ) - STEAMSHIP_API_KEY = gr.Textbox( - label="Steamship api key:", - placeholder="Key to use image generation", - type="text", - ) - HUGGINGFACE_API_KEY = gr.Textbox( - label="Huggingface api key:", - placeholder="Key to use models in huggingface hub", - type="text", - ) - AMADEUS_ID = gr.Textbox( - label="Amadeus id:", placeholder="Id to use Amadeus", type="text" - ) - AMADEUS_KEY = gr.Textbox( - label="Amadeus key:", placeholder="Key to use Amadeus", type="text" - ) + OPENAI_API_KEY = gr.Textbox(label="OpenAI API KEY:", placeholder="sk-...", type="text") + WOLFRAMALPH_APP_ID = gr.Textbox(label="Wolframalpha app id:", placeholder="Key to use wlframalpha", type="text") + WEATHER_API_KEYS = gr.Textbox(label="Weather api key:", placeholder="Key to use weather api", type="text") + BING_SUBSCRIPT_KEY = gr.Textbox(label="Bing subscript key:", placeholder="Key to use bing search", type="text") + ALPHA_VANTAGE_KEY = gr.Textbox(label="Stock api key:", placeholder="Key to use stock api", type="text") + BING_MAP_KEY = gr.Textbox(label="Bing map key:", placeholder="Key to use bing map", type="text") + BAIDU_TRANSLATE_KEY = gr.Textbox(label="Baidu translation key:", placeholder="Key to use baidu translation", type="text") + RAPIDAPI_KEY = gr.Textbox(label="Rapidapi key:", placeholder="Key to use zillow, airbnb and job search", type="text") + SERPER_API_KEY = gr.Textbox(label="Serper key:", placeholder="Key to use google serper and google scholar", type="text") + GPLACES_API_KEY = gr.Textbox(label="Google places key:", placeholder="Key to use google places", type="text") + SCENEX_API_KEY = gr.Textbox(label="Scenex api key:", placeholder="Key to use sceneXplain", type="text") + STEAMSHIP_API_KEY = gr.Textbox(label="Steamship api key:", placeholder="Key to use image generation", type="text") + HUGGINGFACE_API_KEY = gr.Textbox(label="Huggingface api key:", placeholder="Key to use models in huggingface hub", type="text") + AMADEUS_ID = gr.Textbox(label="Amadeus id:", placeholder="Id to use Amadeus", type="text") + AMADEUS_KEY = gr.Textbox(label="Amadeus key:", placeholder="Key to use Amadeus", type="text") key_set_btn = gr.Button(value="Set keys!") + with gr.Tab("Chat with Tool"): with gr.Row(): with gr.Column(scale=4): with gr.Row(): with gr.Column(scale=0.85): - txt = gr.Textbox( - show_label=False, - placeholder="Question here. Use Shift+Enter to add new line.", - lines=1, - ).style(container=False) + txt = gr.Textbox(show_label=False, placeholder="Question here. Use Shift+Enter to add new line.", + lines=1).style(container=False) with gr.Column(scale=0.15, min_width=0): buttonChat = gr.Button("Chat") - memory_utilization = gr.Slider( - label="Memory Utilization:", min=0, max=1, step=0.1, default=0.5 - ) - + memory_utilization = gr.Slider(label="Memory Utilization:", min=0, max=1, step=0.1, default=0.5) + chatbot = gr.Chatbot(show_label=False, visible=True).style(height=600) buttonClear = gr.Button("Clear History") buttonStop = gr.Button("Stop", visible=False) @@ -435,22 +331,13 @@ with gr.Blocks() as demo: with gr.Column(scale=4): with gr.Row(): with gr.Column(scale=1): - model_url = gr.Textbox( - label="VLLM Model URL:", - placeholder="URL to download VLLM model from Hugging Face", - type="text", - ) - buttonDownload = gr.Button("Download Model") - buttonDownload.click( - fn=download_model, inputs=[model_url, memory_utilization] - ) + model_url = gr.Textbox(label="VLLM Model URL:", placeholder="URL to download VLLM model from Hugging Face", type="text"); + buttonDownload = gr.Button("Download Model"); + buttonDownload.click(fn=download_model, inputs=[model_url, memory_utilization]); model_chosen = gr.Dropdown( - list(available_models), - value=DEFAULTMODEL, - multiselect=False, - label="Model provided", - info="Choose the model to solve your question, Default means ChatGPT.", - ) + list(available_models), value=DEFAULTMODEL, multiselect=False, label="Model provided", + info="Choose the model to solve your question, Default means ChatGPT." + ) with gr.Row(): tools_search = gr.Textbox( lines=1, @@ -465,58 +352,38 @@ with gr.Blocks() as demo: info="Choose the tools to solve your question.", ) - key_set_btn.click( - fn=set_environ, - inputs=[ - OPENAI_API_KEY, - WOLFRAMALPH_APP_ID, - WEATHER_API_KEYS, - BING_SUBSCRIPT_KEY, - ALPHA_VANTAGE_KEY, - BING_MAP_KEY, - BAIDU_TRANSLATE_KEY, - RAPIDAPI_KEY, - SERPER_API_KEY, - GPLACES_API_KEY, - SCENEX_API_KEY, - STEAMSHIP_API_KEY, - HUGGINGFACE_API_KEY, - AMADEUS_ID, - AMADEUS_KEY, - ], - outputs=key_set_btn, - ) + key_set_btn.click(fn=set_environ, inputs=[ + OPENAI_API_KEY, + WOLFRAMALPH_APP_ID, + WEATHER_API_KEYS, + BING_SUBSCRIPT_KEY, + ALPHA_VANTAGE_KEY, + BING_MAP_KEY, + BAIDU_TRANSLATE_KEY, + RAPIDAPI_KEY, + SERPER_API_KEY, + GPLACES_API_KEY, + SCENEX_API_KEY, + STEAMSHIP_API_KEY, + HUGGINGFACE_API_KEY, + AMADEUS_ID, + AMADEUS_KEY, + ], outputs=key_set_btn) key_set_btn.click(fn=load_tools, outputs=tools_chosen) tools_search.change(retrieve, tools_search, tools_chosen) buttonSearch.click(clear_retrieve, [], [tools_search, tools_chosen]) - txt.submit( - lambda: [ - gr.update(value=""), - gr.update(visible=False), - gr.update(visible=True), - ], - [], - [txt, buttonClear, buttonStop], - ) - inference_event = txt.submit( - answer_by_tools, - [txt, tools_chosen, model_chosen], - [chatbot, buttonClear, buttonStop], - ) - buttonChat.click( - answer_by_tools, - [txt, tools_chosen, model_chosen], - [chatbot, buttonClear, buttonStop], - ) - buttonStop.click( - lambda: [gr.update(visible=True), gr.update(visible=False)], - [], - [buttonClear, buttonStop], - cancels=[inference_event], - ) + txt.submit(lambda: [gr.update(value=''), gr.update(visible=False), gr.update(visible=True)], [], + [txt, buttonClear, buttonStop]) + inference_event = txt.submit(answer_by_tools, [txt, tools_chosen, model_chosen], [chatbot, buttonClear, buttonStop]) + buttonChat.click(answer_by_tools, [txt, tools_chosen, model_chosen], [chatbot, buttonClear, buttonStop]) + buttonStop.click(lambda: [gr.update(visible=True), gr.update(visible=False)], [], [buttonClear, buttonStop], + cancels=[inference_event]) buttonClear.click(clear_history, [], chatbot) # demo.queue().launch(share=False, inbrowser=True, server_name="127.0.0.1", server_port=7001) demo.queue().launch() + + +