Skip to content

Commit 956d9b9

Browse files
committed
self-hosted: support o3-mini and deprecate o1-preview
1 parent 2fa056c commit 956d9b9

File tree

4 files changed

+16
-8
lines changed

4 files changed

+16
-8
lines changed

docs/self-hosted/azure-devops.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -61,10 +61,12 @@ LLM_PROVIDER=azure-openai
6161
LLM_TIMEOUT=360000
6262
AZURE_OPENAI_ENDPOINT=<azure-openai-endpoint>
6363
AZURE_OPENAI_API_KEY=<key>
64-
## it is recommended to use gpt-4o-mini, o1-mini, and o1-preview deployments. The production release of o1 model is inferior to the preview release as of now. Also, please make sure that the deployment name of o1-preview mentions "o1-preview" in it.
64+
## it is recommended to use gpt-4o-mini, o3-mini, and o1 deployments.
6565
AZURE_GPT4OMINI_DEPLOYMENT_NAME=<gpt-4o-mini-deployment-name>
66-
AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
66+
AZURE_O3MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
6767
AZURE_O1_DEPLOYMENT_NAME=[<o1-deployment-name>]
68+
# optionally, you can swap o3-mini with o1-mini
69+
AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
6870

6971
# if using AWS Bedrock
7072
LLM_PROVIDER=bedrock-anthropic

docs/self-hosted/bitbucket.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -54,10 +54,12 @@ LLM_PROVIDER=azure-openai
5454
LLM_TIMEOUT=360000
5555
AZURE_OPENAI_ENDPOINT=<azure-openai-endpoint>
5656
AZURE_OPENAI_API_KEY=<key>
57-
## it is recommended to use gpt-4o-mini, o1-mini, and o1-preview deployments. The production release of o1 model is inferior to the preview release as of now. Also, please make sure that the deployment name of o1-preview mentions "o1-preview" in it.
57+
## it is recommended to use gpt-4o-mini, o3-mini, and o1 deployments.
5858
AZURE_GPT4OMINI_DEPLOYMENT_NAME=<gpt-4o-mini-deployment-name>
59-
AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
59+
AZURE_O3MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
6060
AZURE_O1_DEPLOYMENT_NAME=[<o1-deployment-name>]
61+
# optionally, you can swap o3-mini with o1-mini
62+
AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
6163

6264
# if using AWS Bedrock
6365
LLM_PROVIDER=bedrock-anthropic

docs/self-hosted/github.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -65,10 +65,12 @@ LLM_PROVIDER=azure-openai
6565
LLM_TIMEOUT=360000
6666
AZURE_OPENAI_ENDPOINT=<azure-openai-endpoint>
6767
AZURE_OPENAI_API_KEY=<key>
68-
## it is recommended to use gpt-4o-mini, o1-mini, and o1-preview deployments. The production release of o1 model is inferior to the preview release as of now. Also, please make sure that the deployment name of o1-preview mentions "o1-preview" in it.
68+
## it is recommended to use gpt-4o-mini, o3-mini, and o1 deployments.
6969
AZURE_GPT4OMINI_DEPLOYMENT_NAME=<gpt-4o-mini-deployment-name>
70-
AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
70+
AZURE_O3MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
7171
AZURE_O1_DEPLOYMENT_NAME=[<o1-deployment-name>]
72+
# optionally, you can swap o3-mini with o1-mini
73+
AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
7274

7375
# if using AWS Bedrock
7476
LLM_PROVIDER=bedrock-anthropic

docs/self-hosted/gitlab.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -60,10 +60,12 @@ LLM_PROVIDER=azure-openai
6060
LLM_TIMEOUT=360000
6161
AZURE_OPENAI_ENDPOINT=<azure-openai-endpoint>
6262
AZURE_OPENAI_API_KEY=<key>
63-
## it is recommended to use gpt-4o-mini, o1-mini, and o1-preview deployments. The production release of o1 model is inferior to the preview release as of now. Also, please make sure that the deployment name of o1-preview mentions "o1-preview" in it.
63+
## it is recommended to use gpt-4o-mini, o3-mini, and o1 deployments.
6464
AZURE_GPT4OMINI_DEPLOYMENT_NAME=<gpt-4o-mini-deployment-name>
65-
AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
65+
AZURE_O3MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
6666
AZURE_O1_DEPLOYMENT_NAME=[<o1-deployment-name>]
67+
# optionally, you can swap o3-mini with o1-mini
68+
AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
6769

6870
# if using AWS Bedrock
6971
LLM_PROVIDER=bedrock-anthropic

0 commit comments

Comments
 (0)
pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy