From b59fa543b10a8b5d7aae7e26c6465584257185ae Mon Sep 17 00:00:00 2001
From: Hasit Mistry <hasit@coderabbit.ai>
Date: Wed, 26 Feb 2025 22:47:38 -0800
Subject: [PATCH] Update OAuth2 configuration with additional parameters and
 links

---
 docs/self-hosted/azure-devops.md | 7 +++++++
 docs/self-hosted/bitbucket.md    | 7 +++++++
 docs/self-hosted/github.md       | 7 +++++++
 docs/self-hosted/gitlab.md       | 7 +++++++
 4 files changed, 28 insertions(+)

diff --git a/docs/self-hosted/azure-devops.md b/docs/self-hosted/azure-devops.md
index 48d8df42..a96c82df 100644
--- a/docs/self-hosted/azure-devops.md
+++ b/docs/self-hosted/azure-devops.md
@@ -72,6 +72,7 @@ AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
 # OAuth2 Configuration (optional)
 # This will use client_credentials flow to get an access token,
 # and use it to make requests to the LLM provider.
+# Here is more information on this flow: https://learn.microsoft.com/en-us/entra/identity-platform/v2-oauth2-client-creds-grant-flow#first-case-access-token-request-with-a-shared-secret
 # It is expected that the response from the OAuth2 server will be in the format
 # {
 #   "access_token": "<access-token>",
@@ -81,6 +82,9 @@ AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
 OAUTH2_ENDPOINT=[<endpoint>]
 OAUTH2_CLIENT_ID=[<client-id>]
 OAUTH2_CLIENT_SECRET=[<client-secret>]
+OAUTH2_SCOPE=[<oauth2-scope>]
+
+HTTPS_PROXY=[<https-proxy-url>]
 
 # if using AWS Bedrock
 LLM_PROVIDER=bedrock-anthropic
@@ -106,6 +110,9 @@ JIRA_HOST=[<jira-host-url>]
 JIRA_PAT=[<jira-personal-access-token>]
 
 LINEAR_PAT=[<linear-personal-access-token>]
+
+ENABLE_WEB_SEARCH=[true]
+PERPLEXITY_API_KEY=[<perplexity-api-key>]
 ```
 
 :::note
diff --git a/docs/self-hosted/bitbucket.md b/docs/self-hosted/bitbucket.md
index f81709c3..74d4c9ae 100644
--- a/docs/self-hosted/bitbucket.md
+++ b/docs/self-hosted/bitbucket.md
@@ -64,6 +64,7 @@ AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
 # OAuth2 Configuration (optional)
 # This will use client_credentials flow to get an access token,
 # and use it to make requests to the LLM provider.
+# Here is more information on this flow: https://learn.microsoft.com/en-us/entra/identity-platform/v2-oauth2-client-creds-grant-flow#first-case-access-token-request-with-a-shared-secret
 # It is expected that the response from the OAuth2 server will be in the format
 # {
 #   "access_token": "<access-token>",
@@ -73,6 +74,9 @@ AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
 OAUTH2_ENDPOINT=[<endpoint>]
 OAUTH2_CLIENT_ID=[<client-id>]
 OAUTH2_CLIENT_SECRET=[<client-secret>]
+OAUTH2_SCOPE=[<oauth2-scope>]
+
+HTTPS_PROXY=[<https-proxy-url>]
 
 # if using AWS Bedrock
 LLM_PROVIDER=bedrock-anthropic
@@ -101,6 +105,9 @@ JIRA_HOST=[<jira-host-url>]
 JIRA_PAT=[<jira-personal-access-token>]
 
 LINEAR_PAT=[<linear-personal-access-token>]
+
+ENABLE_WEB_SEARCH=[true]
+PERPLEXITY_API_KEY=[<perplexity-api-key>]
 ```
 
 :::note
diff --git a/docs/self-hosted/github.md b/docs/self-hosted/github.md
index 18cef2b1..6d081fe5 100644
--- a/docs/self-hosted/github.md
+++ b/docs/self-hosted/github.md
@@ -75,6 +75,7 @@ AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
 # OAuth2 Configuration (optional)
 # This will use client_credentials flow to get an access token,
 # and use it to make requests to the LLM provider.
+# Here is more information on this flow: https://learn.microsoft.com/en-us/entra/identity-platform/v2-oauth2-client-creds-grant-flow#first-case-access-token-request-with-a-shared-secret
 # It is expected that the response from the OAuth2 server will be in the format
 # {
 #   "access_token": "<access-token>",
@@ -84,6 +85,9 @@ AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
 OAUTH2_ENDPOINT=[<endpoint>]
 OAUTH2_CLIENT_ID=[<client-id>]
 OAUTH2_CLIENT_SECRET=[<client-secret>]
+OAUTH2_SCOPE=[<oauth2-scope>]
+
+HTTPS_PROXY=[<https-proxy-url>]
 
 # if using AWS Bedrock
 LLM_PROVIDER=bedrock-anthropic
@@ -119,6 +123,9 @@ JIRA_HOST=[<jira-host-url>]
 JIRA_PAT=[<jira-personal-access-token>]
 
 LINEAR_PAT=[<linear-personal-access-token>]
+
+ENABLE_WEB_SEARCH=[true]
+PERPLEXITY_API_KEY=[<perplexity-api-key>]
 ```
 
 :::note
diff --git a/docs/self-hosted/gitlab.md b/docs/self-hosted/gitlab.md
index 654cef40..44e1f6a1 100644
--- a/docs/self-hosted/gitlab.md
+++ b/docs/self-hosted/gitlab.md
@@ -70,6 +70,7 @@ AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
 # OAuth2 Configuration (optional)
 # This will use client_credentials flow to get an access token,
 # and use it to make requests to the LLM provider.
+# Here is more information on this flow: https://learn.microsoft.com/en-us/entra/identity-platform/v2-oauth2-client-creds-grant-flow#first-case-access-token-request-with-a-shared-secret
 # It is expected that the response from the OAuth2 server will be in the format
 # {
 #   "access_token": "<access-token>",
@@ -79,6 +80,9 @@ AZURE_O1MINI_DEPLOYMENT_NAME=[<o1-mini-deployment-name>]
 OAUTH2_ENDPOINT=[<endpoint>]
 OAUTH2_CLIENT_ID=[<client-id>]
 OAUTH2_CLIENT_SECRET=[<client-secret>]
+OAUTH2_SCOPE=[<oauth2-scope>]
+
+HTTPS_PROXY=[<https-proxy-url>]
 
 # if using AWS Bedrock
 LLM_PROVIDER=bedrock-anthropic
@@ -110,6 +114,9 @@ JIRA_HOST=[<jira-host-url>]
 JIRA_PAT=[<jira-personal-access-token>]
 
 LINEAR_PAT=[<linear-personal-access-token>]
+
+ENABLE_WEB_SEARCH=[true]
+PERPLEXITY_API_KEY=[<perplexity-api-key>]
 ```
 
 :::note