--- title: "Setup" description: "Get your 01 server up and running" --- ## Run Server ```bash poetry run 01 --server ``` ## Configure A core part of the 01 server is the interpreter which is an instance of Open Interpreter. Open Interpreter is highly configurable and only requires updating a single file. ```bash # Edit i.py software/source/server/i.py ``` Properties such as `model`, `context_window`, and many more can be updated here. ### LLM service provider If you wish to use a local model, you can use the `--llm-service` flag: ```bash # use llamafile poetry run 01 --server --llm-service llamafile ``` For more information about LLM service providers, check out the page on Language Models. ### Voice Interface Both speech-to-text and text-to-speech can be configured in 01OS. You are able to pass CLI flags `--tts-service` and/or `--stt-service` with the desired service provider to swap out different services These different service providers can be found in `/services/stt` and `/services/tts` For more information, please read about speech-to-text and text-to-speech ## CLI Flags - `--server` Run server. - `--server-host TEXT` Specify the server host where the server will deploy. Default: `0.0.0.0`. - `--server-port INTEGER` Specify the server port where the server will deploy. Default: `10001`. - `--tunnel-service TEXT` Specify the tunnel service. Default: `ngrok`. - `--expose` Expose server to internet. - `--server-url TEXT` Specify the server URL that the client should expect. Defaults to server-host and server-port. Default: `None`. - `--llm-service TEXT` Specify the LLM service. Default: `litellm`. - `--model TEXT` Specify the model. Default: `gpt-4`. - `--llm-supports-vision` Specify if the LLM service supports vision. - `--llm-supports-functions` Specify if the LLM service supports functions. - `--context-window INTEGER` Specify the context window size. Default: `2048`. - `--max-tokens INTEGER` Specify the maximum number of tokens. Default: `4096`. - `--temperature FLOAT` Specify the temperature for generation. Default: `0.8`. - `--tts-service TEXT` Specify the TTS service. Default: `openai`. - `--stt-service TEXT` Specify the STT service. Default: `openai`. - `--local` Use recommended local services for LLM, STT, and TTS. - `--install-completion [bash|zsh|fish|powershell|pwsh]` Install completion for the specified shell. Default: `None`. - `--show-completion [bash|zsh|fish|powershell|pwsh]` Show completion for the specified shell, to copy it or customize the installation. Default: `None`. - `--help` Show this message and exit.