This repository shows how you can use Bicep to deploy flagship models in Azure AI Foundry.
This project framework provides the following features:
- Deploy multiple models from the Azure AI catalog into the Azure AI model inference service in Azure AI Services.
- Deploy multiple models in Azure AI Services and provision/configure a project to run inference from the Azure AI model inference service in Azure AI Services.
- Configure Content filters for the model deployments.
-
Download and install it from the Azure CLI installation page.
-
Verify the installation:
az --version
-
An Azure subscription.
-
Sign in to your Azure account to deploy resources.
az login
-
The
jq
tool.sudo apt install jq
Clone the project:
git clone https://github.com/Azure-Samples/azureai-model-inference-bicep`
cd azureai-model-inference-bicep/infra
Inspect the file models.json
and configure the models you are interested on deploying. The file is an array of JSON objects with the properties provider
, name
, and version. By default, this repository contains all the model definitions available for pay-as-you-go:
{
"provider": "Microsoft",
"name": "Phi-3.5-vision-instruct",
"version": "2",
"sku": "GlobalStandard"
}
Ensure you are in the right subscription:
az account set --subscription "<subscription-id>"
Configure the deployment
RESOURCE_GROUP="azureai-models-dev"
LOCATION="eastus2"
Run the deployment:
az deployment group create \
--resource-group $RESOURCE_GROUP \
--template-file deploy.bicep
Run the deployment:
az deployment group create \
--resource-group $RESOURCE_GROUP \
--template-file deploy-with-project.bicep