Explore image and video capable models in Azure AI Foundry

Azure AI Foundry includes a wide range of generative AI models, some of which are capable of computer vision related tasks. In this exercise, you’ll use models to analyze and generate visual content.

Analyze images

While large language models (LLMs) are most commonly associated with responding to text-based prompts, many newer models have multi-modal capabilities that enable them to process image-based input.

  1. In a web browser, open Azure AI Foundry portal at https://ai.azure.com and sign in using your Azure credentials. Close any tips or quick start panes that are opened the first time you sign in, and if necessary use the Azure AI Foundry logo at the top left to navigate to the home page, which looks similar to the following image (close the Help pane if it’s open):

    Screenshot of Azure AI Foundry portal.

  2. Review the information on the home page.

  3. In the home page, in the Explore models and capabilities section, search for the gpt-4o-mini model; which we’ll use in our project.

  4. In the search results, select the gpt-4o-mini model to see its details, and then at the top of the page for the model, select Use this model.

  5. When prompted to create a project, enter a valid name for your project and expand Advanced options.

  6. Select Customize and specify the following settings for your hub:
    • Azure AI Foundry resource: A valid name for your Azure AI Foundry resource
    • Subscription: Your Azure subscription
    • Resource group: Create or select a resource group
    • Region: Select a recommended region*

    * Some Azure AI resources are constrained by regional model quotas. In the event of a quota limit being exceeded later in the exercise, there’s a possibility you may need to create another resource in a different region.

  7. Select Create and wait for your project, including the gpt-4o-mini model deployment you selected, to be created.

    Note: Depending on your available quota for gpt-4o-mini models you might receive a additional prompt to deploy the model to a resource in a different region. If this happens, do so, using the default settings.

  8. When your project is created, your deployed model will be displayed:

    Screenshot of the model deployment page.

    Now you can test your multimodal model deployment with an image-based prompt in the chat playground.

  9. In the navigation pane on the left, select the Playgrounds page, and then open the Chat playground.

  10. In a new browser tab, download image.zip from https://github.com/GraemeMalcolm/openai-vision/raw/refs/heads/main/Labfiles/images/image.zip and extract it in a folder on your local file system.

  11. On the chat playground page, in the Setup pane, ensure that your gpt-4o-mini model model deployment is selected.

  12. In the main chat session panel, under the chat input box, use the attach button (📎) to upload any of the extracted image files, and then add the text What animal is this? and submit the prompt.

    Screenshot of the chat playground page.

  13. Review the response, which should hopefully provide relevant information about the animal in the image.
  14. Attach a different image and enter the following prompt:

    Suggest a caption and some tags for this image.
    
  15. Continue to experiment, uploading images and prompting the model to interpret them.

Generate images

Some LLMs are designed to generate visual responses, based on natural language descriptions of a desired image.

  1. In the pane on the left for your project, in the My assets section, select the Models + endpoints page.
  2. In the Models + endpoints page, in the Model deployments tab, in the + Deploy model menu, select Deploy base model.
  3. Search for the dall-e-3 model in the list, and then select and confirm it.
  4. Deploy the model using the default settings.

    Note: Depending on your available quota for DALL-E-3 models you might receive a additional prompt to deploy the model to a resource in a different region. If this happens, do so, using the default settings.

  5. Wait for the deployment provisioning state to be Completed.

  6. In the navigation pane on the left, select the Playgrounds page, and then open the Images playground.
  7. Ensure your DALL-E model deployment is selected. Then, in the Prompt box, enter a prompt such as A photorealistic giraffe walking in the Serengeti.
  8. Review the resulting image in the playground:

    Screenshot of the images playground with a generated image.

  9. Enter another prompt, such as A safari truck with tourists viewing a herd of zebra and review the resulting image.
  10. Continue testing with new prompts to create images.

Generate video

SIn addition to static images, you can use some models to generate video.

  1. In the pane on the left for your project, in the My assets section, select the Models + endpoints page.
  2. In the Models + endpoints page, in the Model deployments tab, in the + Deploy model menu, select Deploy base model.
  3. Search for the sora model in the list, and then select and confirm it.
  4. Deploy the model using the default settings.

    Note: Depending on your available quota for Sora models you might receive a additional prompt to deploy the model to a resource in a different region. If this happens, do so, using the default settings.

  5. Wait for the deployment provisioning state to be Completed.

  6. In the navigation pane on the left, select the Playgrounds page, and then open the Video playground.
  7. Ensure your Sora model deployment is selected. Then, in the Prompt box, enter a prompt such as A lion, lying by a tree in the savannah and roaring.
  8. Wait for the video to be generated, then view it in the playground:

    Screenshot of the video playground with a generated image.

  9. Close the video, and then enter another prompt, such as A herd of wildebeest migrating across a river and review the resulting video.

Clean up

If you’ve finished exploring models work working with images and video in Azure AI Foundry, you should delete the resources you have created in this exercise to avoid incurring unnecessary Azure costs.

  1. Open the Azure portal at https://portal.azure.com and view the contents of the resource group where you deployed the resources used in this exercise.
  2. On the toolbar, select Delete resource group.
  3. Enter the resource group name and confirm that you want to delete it.