update model

pull/570/head
Richard Anthony Hein 8 months ago
parent d30f5f8259
commit 85087aecd8

@ -12,7 +12,7 @@ weather_api_key= "af6ef989b5c50a91ca068cc00df125b7", # Replace with your weathe
llm = OpenAIChatLLM( llm = OpenAIChatLLM(
base_url=api_base, base_url=api_base,
api_key=api_key, api_key=api_key,
model="NousResearch/Meta-Llama-3-8B-Instruct", model="NousResearch/Meta-Llama-3.1-8B-Instruct",
temperature=0, temperature=0,
streaming=False streaming=False
) )

@ -1 +1 @@
Subproject commit 9d23e948f0a64c2e8aa9cdb6a5020222f7b4d325 Subproject commit bc2223311c531b40b3b8e42cac743eda325947e6

@ -1,5 +1 @@
from playground.demos.chatbot.server.vector_store import VectorStorage
__all__ = [
"VectorStorage",
]

@ -5,4 +5,4 @@ docker run --runtime nvidia --gpus all \
--network=host \ --network=host \
--name vllm \ --name vllm \
vllm/vllm-openai:latest \ vllm/vllm-openai:latest \
--model NousResearch/Meta-Llama-3-8B-Instruct --model NousResearch/Meta-Llama-3.1-8B-Instruct

@ -2,20 +2,21 @@
import logging import logging
import os import os
from urllib.parse import urlparse from urllib.parse import urlparse
from swarms.structs.agent import Agent
from dotenv import load_dotenv from dotenv import load_dotenv
from fastapi import FastAPI, HTTPException, Request from fastapi import FastAPI, HTTPException, Request
from fastapi.middleware.cors import CORSMiddleware from fastapi.middleware.cors import CORSMiddleware
from fastapi.responses import JSONResponse from fastapi.responses import JSONResponse
from fastapi.routing import APIRouter from fastapi.routing import APIRouter
from fastapi.staticfiles import StaticFiles from fastapi.staticfiles import StaticFiles
from huggingface_hub import login
from swarms.prompts.chat_prompt import Message, Role
from swarms.prompts.conversational_RAG import QA_PROMPT_TEMPLATE_STR
from playground.demos.chatbot.server.responses import StreamingResponse from playground.demos.chatbot.server.responses import StreamingResponse
from playground.demos.chatbot.server.server_models import ChatRequest from playground.demos.chatbot.server.server_models import ChatRequest
from playground.demos.chatbot.server.vector_storage import RedisVectorStorage from playground.demos.chatbot.server.vector_storage import RedisVectorStorage
from swarms.models.popular_llms import OpenAIChatLLM from swarms.models.popular_llms import OpenAIChatLLM
# from huggingface_hub import login
from swarms.prompts.chat_prompt import Message, Role
from swarms.prompts.conversational_RAG import QA_PROMPT_TEMPLATE_STR
from swarms.structs.agent import Agent
logging.basicConfig(level=logging.ERROR) logging.basicConfig(level=logging.ERROR)
@ -26,9 +27,9 @@ dotenv_path = os.path.join(
) )
load_dotenv(dotenv_path) load_dotenv(dotenv_path)
hf_token = os.environ.get( # hf_token = os.environ.get(
"HUGGINFACEHUB_API_KEY" # "HUGGINFACEHUB_API_KEY"
) # Get the Huggingface API Token # ) # Get the Huggingface API Token
uploads = os.environ.get( uploads = os.environ.get(
"UPLOADS" "UPLOADS"
@ -47,7 +48,7 @@ openai_api_base = (
) )
env_vars = [ env_vars = [
hf_token, # hf_token,
uploads, uploads,
openai_api_key, openai_api_key,
openai_api_base, openai_api_base,
@ -70,7 +71,7 @@ print(f"USE_GPU={use_gpu}")
print(f"OPENAI_API_KEY={openai_api_key}") print(f"OPENAI_API_KEY={openai_api_key}")
print(f"OPENAI_API_BASE={openai_api_base}") print(f"OPENAI_API_BASE={openai_api_base}")
print("Logging in to huggingface.co...") print("Logging in to huggingface.co...")
login(token=hf_token) # login to huggingface.co # login(token=hf_token) # login to huggingface.co
app = FastAPI(title="Chatbot") app = FastAPI(title="Chatbot")
@ -176,7 +177,7 @@ async def create_chat(
# add docs to short term memory # add docs to short term memory
for data in [doc["content"] for doc in docs]: for data in [doc["content"] for doc in docs]:
agent.add_message_to_memory(role=Role.HUMAN, content=data) agent.add_message_to_memory(data)
async for response in agent.run_async(messages[-1].content): async for response in agent.run_async(messages[-1].content):
res = response res = response

@ -44,8 +44,8 @@ class ChatRequest(BaseModel):
""" The model for a ChatRequest for theChatbot Chat POST endpoint""" """ The model for a ChatRequest for theChatbot Chat POST endpoint"""
id: str id: str
model: AIModel = AIModel( model: AIModel = AIModel(
id="NousResearch/Meta-Llama-3-8B-Instruct", id="NousResearch/Meta-Llama-3.1-8B-Instruct",
name="NousResearch/Meta-Llama-3-8B-Instruct", name="NousResearch/Meta-Llama-3.1-8B-Instruct",
maxLength=2048, maxLength=2048,
tokenLimit=2048, tokenLimit=2048,
) )

@ -11,7 +11,7 @@ api_base = os.getenv("OPENAI_API_BASE") or "http://localhost:8000/v1" # for vllm
# Create an instance of the OpenAIChat class # Create an instance of the OpenAIChat class
model = OpenAIChat( model = OpenAIChat(
base_url=api_base, api_key=api_key, model="NousResearch/Meta-Llama-3-8B-Instruct", temperature=0.5, streaming=True, verbose=True base_url=api_base, api_key=api_key, model="NousResearch/Meta-Llama-3.1-8B-Instruct", temperature=0.5, streaming=True, verbose=True
) )
# Initialize the agent # Initialize the agent

Loading…
Cancel
Save