@@ -6,16 +6,22 @@ targetScope = 'resourceGroup'
66@description ('A unique prefix for all resources in this deployment. This should be 3-20 characters long:' )
77param environmentName string
88
9+ @minLength (1 )
10+ @description ('Secondary location for databases creation(example:eastus2):' )
11+ param secondaryLocation string
12+
913@metadata ({
1014 azd : {
1115 type : 'location'
16+ usageName : [
17+ 'OpenAI.GlobalStandard.gpt4.1,200'
18+ 'OpenAI.Standard.text-embedding-ada-002,80'
19+ ]
1220 }
1321})
1422
23+ param AZURE_AI_SERVICE_LOCATION string
1524
16- @minLength (1 )
17- @description ('Secondary location for databases creation(example:eastus2):' )
18- param secondaryLocation string
1925
2026@minLength (1 )
2127@description ('GPT model deployment type:' )
@@ -37,7 +43,7 @@ param azureOpenaiAPIVersion string = '2024-05-01-preview'
3743@description ('Capacity of the GPT deployment:' )
3844// You can increase this, but capacity is limited per model/region, so you will get errors if you go over
3945// https://learn.microsoft.com/en-us/azure/ai-services/openai/quotas-limits
40- param gptDeploymentCapacity int = 30
46+ param gptDeploymentCapacity int = 200
4147
4248@minLength (1 )
4349@description ('Name of the Text Embedding model to deploy:' )
@@ -63,6 +69,8 @@ var uniqueId = toLower(uniqueString(environmentName, subscription().id, solution
6369var solutionPrefix = 'dg${padLeft (take (uniqueId , 12 ), 12 , '0' )}'
6470
6571
72+
73+
6674// ========== Managed Identity ========== //
6775module managedIdentityModule 'deploy_managed_identity.bicep' = {
6876 name : 'deploy_managed_identity'
@@ -91,7 +99,7 @@ module aifoundry 'deploy_ai_foundry.bicep' = {
9199 name : 'deploy_ai_foundry'
92100 params : {
93101 solutionName : solutionPrefix
94- solutionLocation : solutionLocation
102+ solutionLocation : AZURE_AI_SERVICE_LOCATION
95103 keyVaultName : kvault .outputs .keyvaultName
96104 deploymentType : deploymentType
97105 gptModelName : gptModelName
0 commit comments