Exploring Evaluation in Azure AI Foundry: A Deep Dive

Published On Thu Dec 26 2024
Exploring Evaluation in Azure AI Foundry: A Deep Dive

TomazTsql | Tomaz doing BI and DEV with SQL Server and R ...

In this Microsoft Azure AI series:

Documentation and Resources for Azure AI Foundry

Documentation and material for Azure AI Foundry are plentiful and growing on a daily basis, since the topic on AI and GenAI is evermore so popular. The general and best way to start is with Microsoft own product introduction website: https://azure.microsoft.com/en-us/products/ai-studio and later you can scale with different material from many different websites.

Microsoft Ignite

Detailed product documentation is on Microsoft learn: https://learn.microsoft.com/en-us/azure/ai-studio/

Video series:
Learning paths:
Books:
Code and samples:

Many introduction articles are available as well as many listings for the use cases. All of the code samples will be available on my Github.

In this Microsoft Azure AI series:

Evaluation in Azure AI Foundry

With evaluation you performing iterative, systematic evaluations with the right evaluators and measure and address potential response quality, safety, or security concerns throughout the AI development lifecycle, from initial model selection through post-production monitoring. With the Evaluation in Azure AI Foundry, you can evaluate the GenAI Ops Lifecycle production. In addition, it also gives you the ability to assess the frequency and severity of content risks or undesirable behavior in AI responses.

Evaluation of generative AI applications with Azure AI Foundry

With each model, you can consider and evaluate:

  • Evaluation of the models
  • Evaluate different metrics

And you can start with evaluation when the deployment is ready, all relevant metrics selected, prompts added and data integration (input and output) defined. The results of the evaluation can be analyzed with the Python SDK as well.

Tomorrow we will look into documentation for Azure AI Foundry. All of the code samples will be available on my Github.

In this Microsoft Azure AI series:

Tracing with Azure AI Inference SDK

Tracing is a powerful tool that offers developers an in-depth understanding of the execution process of their generative AI applications. Though still in preview (in the time of writing this post), It provides a detailed view of the execution flow of the application and the essential information for debugging or optimisations.

Tracing with the Azure AI Inference SDK offers enhanced visibility and simplified troubleshooting for LLM-based applications, effectively supporting development, iteration, and production monitoring. Tracing follows the OpenTelemetry semantic conventions, capturing and visualizing the internal execution details of any AI application, enhancing the overall development experience.

How to export and import a fine-tuned AI model to a different ...

Key advances of using tracing are:

  • Creating a new app or connecting to an existing one
  • Using the Kusto language for analyzing the logs for attributes, content, spans, evaluation events, and others
  • Collecting and analyzing log traces with Python SDK

Tomorrow we will look into Evaluation in Azure AI Foundry. All of the code samples will be available on my Github.

In this Microsoft Azure AI series:

Prompt Flow in Azure AI Foundry

Prompt Flow is particularly beneficial for organizations leveraging AI to streamline operations, enhance customer experiences, and innovate in digital transformation projects.

With Python, you can start using prompt flow by installing the package and create a prompty file. For additional information on Prompt flow, visit: https://microsoft.github.io/promptflow/index.html

The main building blocks of Prompt flow are:

  • Creating standard flow, chat flow, or evaluation flow
  • Web Classification flow example

You can start building flow with the three predefined types or clone the prepared flows and deploy it based on your needs.

Tomorrow we will look into building prompt flow in VS Code using Python. All of the code samples will be available on my Github.

In this Microsoft Azure AI series:

Model Deployment in Azure AI Foundry

Models from the model catalog can be deployed using programming languages or using the Foundry studio. Model deployment has two types: Deploy from the base model or deploy from the fine-tuned model.

Each model has all the necessary detail information, as well all the metrics to check number of requests, prompt token counts, number of requests and additional information on model usage.

Models from the model catalog can also be deployed as a serverless API with pay-as-you-go billing. Deployment with Python SDK is straightforward.

Tomorrow we will look into the prompt flow in Azure AI Foundry. All of the code samples will be available on my Github.