Skip to content

Describes how to deploy Azure AI model inference service in Azure AI Services with Bicep

License

Notifications You must be signed in to change notification settings

Azure-Samples/azureai-model-inference-bicep

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Deploy and configure model inference in Azure AI Foundry

This repository shows how you can use Bicep to deploy flagship models in Azure AI Foundry.

Features

This project framework provides the following features:

  • Deploy multiple models from the Azure AI catalog into the Azure AI model inference service in Azure AI Services.
  • Deploy multiple models in Azure AI Services and provision/configure a project to run inference from the Azure AI model inference service in Azure AI Services.
  • Configure Content filters for the model deployments.

Getting Started

Prerequisites

  • Download and install it from the Azure CLI installation page.

  • Verify the installation:

    az --version
  • An Azure subscription.

  • Sign in to your Azure account to deploy resources.

    az login
  • The jq tool.

    sudo apt install jq

Quickstart

Clone the project:

git clone https://github.com/Azure-Samples/azureai-model-inference-bicep`
cd azureai-model-inference-bicep/infra

Inspect the file models.json and configure the models you are interested on deploying. The file is an array of JSON objects with the properties provider, name, and version. By default, this repository contains all the model definitions available for pay-as-you-go:

{
    "provider": "Microsoft",
    "name": "Phi-3.5-vision-instruct",
    "version": "2",
    "sku": "GlobalStandard"
}

Ensure you are in the right subscription:

az account set --subscription "<subscription-id>"

Configure the deployment

RESOURCE_GROUP="azureai-models-dev"
LOCATION="eastus2"

Deploy all the models in Azure AI Services

Run the deployment:

az deployment group create \
  --resource-group $RESOURCE_GROUP \
  --template-file deploy.bicep

Deploy all the models and configure an AI project/hub

Run the deployment:

az deployment group create \
  --resource-group $RESOURCE_GROUP \
  --template-file deploy-with-project.bicep

Resources

About

Describes how to deploy Azure AI model inference service in Azure AI Services with Bicep

Resources

License

Code of conduct

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •