[DEMO][Gemini VCOT]

pull/317/head
Kye 1 year ago
parent 3a9d428ebe
commit b54626f98c

@ -0,0 +1,29 @@
import os
from dotenv import load_dotenv
from swarms.models import Gemini
from swarms.prompts.visual_cot import VISUAL_CHAIN_OF_THOUGHT
# Load the environment variables
load_dotenv()
# Get the API key from the environment
api_key = os.environ.get("GEMINI_API_KEY")
# Initialize the language model
llm = Gemini(
gemini_api_key=api_key,
temperature=0.5,
max_tokens=1000,
system_prompt=VISUAL_CHAIN_OF_THOUGHT
)
# Initialize the task
task = "This is an eye test. What do you see?"
img = "playground/demos/multi_modal_chain_of_thought/eyetest.jpg"
# Run the workflow on a task
out = llm.run(task=task, img=img)
print(out)

@ -49,7 +49,7 @@ class Gemini(BaseMultiModalModel):
stream (bool, optional): _description_. Defaults to False. stream (bool, optional): _description_. Defaults to False.
candidate_count (int, optional): _description_. Defaults to 1. candidate_count (int, optional): _description_. Defaults to 1.
stop_sequence ([type], optional): _description_. Defaults to ['x']. stop_sequence ([type], optional): _description_. Defaults to ['x'].
max_output_tokens (int, optional): _description_. Defaults to 100. max_tokens (int, optional): _description_. Defaults to 100.
temperature (float, optional): _description_. Defaults to 0.9. temperature (float, optional): _description_. Defaults to 0.9.
Methods: Methods:
@ -80,7 +80,7 @@ class Gemini(BaseMultiModalModel):
stream: bool = False, stream: bool = False,
candidate_count: int = 1, candidate_count: int = 1,
stop_sequence=["x"], stop_sequence=["x"],
max_output_tokens: int = 100, max_tokens: int = 100,
temperature: float = 0.9, temperature: float = 0.9,
system_prompt: str = None, system_prompt: str = None,
*args, *args,
@ -94,7 +94,7 @@ class Gemini(BaseMultiModalModel):
self.stream = stream self.stream = stream
self.candidate_count = candidate_count self.candidate_count = candidate_count
self.stop_sequence = stop_sequence self.stop_sequence = stop_sequence
self.max_output_tokens = max_output_tokens self.max_tokens = max_tokens
self.temperature = temperature self.temperature = temperature
self.system_prompt = system_prompt self.system_prompt = system_prompt
@ -102,7 +102,7 @@ class Gemini(BaseMultiModalModel):
self.generation_config = GenerationConfig( self.generation_config = GenerationConfig(
candidate_count=candidate_count, candidate_count=candidate_count,
# stop_sequence=stop_sequence, # stop_sequence=stop_sequence,
max_output_tokens=max_output_tokens, max_tokens=max_tokens,
temperature=temperature, temperature=temperature,
*args, *args,
**kwargs, **kwargs,

Loading…
Cancel
Save