OpenAI
To work with OpenAI’s models, you need to provide your OpenAI API key either in the OPENAI_API_KEY
environment variable or via the --api-key openai=<key>
command line switch.
First, install aider:
python -m pip install aider-install aider-install
Then configure your API keys:
export OPENAI_API_KEY=<key> # Mac/Linux setx OPENAI_API_KEY <key> # Windows, restart shell after setx
Start working with aider and OpenAI on your codebase:
# Change directory into your codebase cd /to/your/project # o3-mini aider --model o3-mini # o1-mini aider --model o1-mini # GPT-4o aider --model gpt-4o # List models available from OpenAI aider --list-models openai/
You can use aider --model <model-name>
to use any other OpenAI model. For example, if you want to use a specific version of GPT-4 Turbo you could do aider --model gpt-4-0125-preview
.
Reasoning models from other providers
Many of OpenAI’s “reasoning” models have restrictions on streaming and setting the temperature parameter. Some also support different levels of “reasoning effort”. Aider is configured to work properly with these models when served through major provider APIs and has a --reasoning-effort
setting.
You may need to configure reasoning model settings if you are using them through another provider and see errors related to temperature or system prompt.