diff --git a/software/source/clients/linux/device.py b/software/source/clients/linux/device.py index cf549d5..0fa0fed 100644 --- a/software/source/clients/linux/device.py +++ b/software/source/clients/linux/device.py @@ -3,9 +3,8 @@ from ..base_device import Device device = Device() -def main(server_url, tts_service): +def main(server_url): device.server_url = server_url - device.tts_service = tts_service device.start() diff --git a/software/source/clients/mac/device.py b/software/source/clients/mac/device.py index cf549d5..0fa0fed 100644 --- a/software/source/clients/mac/device.py +++ b/software/source/clients/mac/device.py @@ -3,9 +3,8 @@ from ..base_device import Device device = Device() -def main(server_url, tts_service): +def main(server_url): device.server_url = server_url - device.tts_service = tts_service device.start() diff --git a/software/source/clients/windows/device.py b/software/source/clients/windows/device.py index cf549d5..0fa0fed 100644 --- a/software/source/clients/windows/device.py +++ b/software/source/clients/windows/device.py @@ -3,9 +3,8 @@ from ..base_device import Device device = Device() -def main(server_url, tts_service): +def main(server_url): device.server_url = server_url - device.tts_service = tts_service device.start() diff --git a/software/start.py b/software/start.py index 5b711da..f93d872 100644 --- a/software/start.py +++ b/software/start.py @@ -7,9 +7,6 @@ import importlib from source.server.tunnel import create_tunnel from source.server.async_server import main -# from source.server.server import main -from source.server.utils.local_mode import select_local_model - import signal app = typer.Typer() @@ -41,64 +38,20 @@ def run( client_type: str = typer.Option( "auto", "--client-type", help="Specify the client type" ), - llm_service: str = typer.Option( - "litellm", "--llm-service", help="Specify the LLM service" - ), - model: str = typer.Option("gpt-4", "--model", help="Specify the model"), - llm_supports_vision: bool = typer.Option( - False, - "--llm-supports-vision", - help="Specify if the LLM service supports vision", - ), - llm_supports_functions: bool = typer.Option( - False, - "--llm-supports-functions", - help="Specify if the LLM service supports functions", - ), - context_window: int = typer.Option( - 2048, "--context-window", help="Specify the context window size" - ), - max_tokens: int = typer.Option( - 4096, "--max-tokens", help="Specify the maximum number of tokens" - ), - temperature: float = typer.Option( - 0.8, "--temperature", help="Specify the temperature for generation" - ), - tts_service: str = typer.Option( - "elevenlabs", "--tts-service", help="Specify the TTS service" - ), - stt_service: str = typer.Option( - "openai", "--stt-service", help="Specify the STT service" - ), - local: bool = typer.Option( - False, "--local", help="Use recommended local services for LLM, STT, and TTS" - ), - qr: bool = typer.Option(False, "--qr", help="Print the QR code for the server URL"), - mobile: bool = typer.Option( - False, "--mobile", help="Toggle server to support mobile app" + qr: bool = typer.Option( + False, "--qr", help="Display QR code to scan to connect to the server" ), ): _run( - server=server or mobile, + server=server, server_host=server_host, server_port=server_port, tunnel_service=tunnel_service, - expose=expose or mobile, + expose=expose, client=client, server_url=server_url, client_type=client_type, - llm_service=llm_service, - model=model, - llm_supports_vision=llm_supports_vision, - llm_supports_functions=llm_supports_functions, - context_window=context_window, - max_tokens=max_tokens, - temperature=temperature, - tts_service=tts_service, - stt_service=stt_service, - local=local, - qr=qr or mobile, - mobile=mobile, + qr=qr, ) @@ -111,24 +64,8 @@ def _run( client: bool = False, server_url: str = None, client_type: str = "auto", - llm_service: str = "litellm", - model: str = "gpt-4", - llm_supports_vision: bool = False, - llm_supports_functions: bool = False, - context_window: int = 2048, - max_tokens: int = 4096, - temperature: float = 0.8, - tts_service: str = "elevenlabs", - stt_service: str = "openai", - local: bool = False, qr: bool = False, - mobile: bool = False, ): - if local: - tts_service = "coqui" - # llm_service = "llamafile" - stt_service = "local-whisper" - select_local_model() system_type = platform.system() if system_type == "Windows": @@ -156,21 +93,10 @@ def _run( main( server_host, server_port, - tts_service, - # llm_service, - # model, - # llm_supports_vision, - # llm_supports_functions, - # context_window, - # max_tokens, - # temperature, - # stt_service, - # mobile, ), ), ) server_thread.start() - print("server thread started") if expose: tunnel_thread = threading.Thread( @@ -199,11 +125,8 @@ def _run( f".clients.{client_type}.device", package="source" ) - client_thread = threading.Thread( - target=module.main, args=[server_url, tts_service] - ) + client_thread = threading.Thread(target=module.main, args=[server_url]) client_thread.start() - print("client thread started") try: if server: