parent
34bd6eabc9
commit
4850b4a9f3
@ -1,31 +1,42 @@
|
|||||||
async def main(server_host, server_port, tts_service, asynchronous):
|
from interpreter import interpreter
|
||||||
|
|
||||||
if asynchronous:
|
# This is an Open Interpreter compatible profile.
|
||||||
|
# Visit https://01.openinterpreter.com/profile for all options.
|
||||||
|
|
||||||
base_interpreter.system_message = (
|
# 01 suports OpenAI, ElevenLabs, and Coqui (Local) TTS providers
|
||||||
"You are a helpful assistant that can answer questions and help with tasks."
|
# {OpenAI: "openai", ElevenLabs: "elevenlabs", Coqui: "coqui"}
|
||||||
)
|
interpreter.tts = "elevenlabs"
|
||||||
|
|
||||||
base_interpreter.computer.import_computer_api = False
|
|
||||||
|
|
||||||
base_interpreter.llm.model = "groq/llama3-8b-8192"
|
|
||||||
|
|
||||||
base_interpreter.llm.api_key = os.environ["GROQ_API_KEY"]
|
|
||||||
|
|
||||||
base_interpreter.llm.supports_functions = False
|
|
||||||
|
|
||||||
base_interpreter.auto_run = True
|
# 01 Language Model Config.
|
||||||
|
interpreter.llm_service = "litellm"
|
||||||
|
interpreter.llm.model = "groq/llama3-8b-8192"
|
||||||
|
interpreter.llm.supports_vision = False
|
||||||
|
interpreter.llm.supports_functions = False
|
||||||
|
interpreter.llm.context_window = 2048
|
||||||
|
interpreter.llm.max_tokens = 4096
|
||||||
|
interpreter.llm.temperature = 0.8
|
||||||
|
|
||||||
base_interpreter.tts = tts_service
|
# interpreter.llm.api_key = os.environ["GROQ_API_KEY"]
|
||||||
|
|
||||||
interpreter = AsyncInterpreter(base_interpreter)
|
interpreter.computer.import_computer_api = False
|
||||||
|
|
||||||
else:
|
interpreter.auto_run = True
|
||||||
|
interpreter.system_message = (
|
||||||
configured_interpreter = configure_interpreter(base_interpreter)
|
"You are a helpful assistant that can answer questions and help with tasks."
|
||||||
|
)
|
||||||
configured_interpreter.llm.supports_functions = True
|
|
||||||
|
# TODO: include other options in comments in the profiles for tts
|
||||||
configured_interpreter.tts = tts_service
|
# direct people to the profiles directory to make changes to the interpreter profile
|
||||||
|
# this should be made explicit on the docs
|
||||||
interpreter = AsyncInterpreter(configured_interpreter)
|
|
||||||
|
"""
|
||||||
|
llm_service: str = "litellm",
|
||||||
|
model: str = "gpt-4",
|
||||||
|
llm_supports_vision: bool = False,
|
||||||
|
llm_supports_functions: bool = False,
|
||||||
|
context_window: int = 2048,
|
||||||
|
max_tokens: int = 4096,
|
||||||
|
temperature: float = 0.8,
|
||||||
|
tts_service: str = "elevenlabs",
|
||||||
|
stt_service: str = "openai",
|
||||||
|
"""
|
||||||
|
Loading…
Reference in new issue