---
title: "Setup"
description: "Get your 01 server up and running"
---
## Run Server
```bash
poetry run 01  --server
```
## Configure
A core part of the 01 server is the interpreter which is an instance of Open Interpreter.
Open Interpreter is highly configurable and only requires updating a single file.
```bash
# Edit i.py
software/source/server/i.py
```
Properties such as `model`, `context_window`, and many more can be updated here.
### LLM service provider
If you wish to use a local model, you can use the `--llm-service` flag:
```bash
# use llamafile
poetry run 01 --server --llm-service llamafile
```
For more information about LLM service providers, check out the page on Language Models.
### Voice Interface
Both speech-to-text and text-to-speech can be configured in 01OS.
You are able to pass CLI flags `--tts-service` and/or `--stt-service` with the desired service provider to swap out different services
These different service providers can be found in `/services/stt` and `/services/tts`
For more information, please read about speech-to-text and text-to-speech
## CLI Flags
- `--server`  
  Run server.
- `--server-host TEXT`  
  Specify the server host where the server will deploy.  
  Default: `0.0.0.0`.
- `--server-port INTEGER`  
  Specify the server port where the server will deploy.  
  Default: `10001`.
- `--tunnel-service TEXT`  
  Specify the tunnel service.  
  Default: `ngrok`.
- `--expose`  
  Expose server to internet.
- `--server-url TEXT`  
  Specify the server URL that the client should expect.  
  Defaults to server-host and server-port.  
  Default: `None`.
- `--llm-service TEXT`  
  Specify the LLM service.  
  Default: `litellm`.
- `--model TEXT`  
  Specify the model.  
  Default: `gpt-4`.
- `--llm-supports-vision`  
  Specify if the LLM service supports vision.
- `--llm-supports-functions`  
  Specify if the LLM service supports functions.
- `--context-window INTEGER`  
  Specify the context window size.  
  Default: `2048`.
- `--max-tokens INTEGER`  
  Specify the maximum number of tokens.  
  Default: `4096`.
- `--temperature FLOAT`  
  Specify the temperature for generation.  
  Default: `0.8`.
- `--tts-service TEXT`  
  Specify the TTS service.  
  Default: `openai`.
- `--stt-service TEXT`  
  Specify the STT service.  
  Default: `openai`.
- `--local`  
  Use recommended local services for LLM, STT, and TTS.
- `--install-completion [bash|zsh|fish|powershell|pwsh]`  
  Install completion for the specified shell.  
  Default: `None`.
- `--show-completion [bash|zsh|fish|powershell|pwsh]`  
  Show completion for the specified shell, to copy it or customize the installation.  
  Default: `None`.
- `--help`  
  Show this message and exit.