Build Generative AI Solutions with Azure AI Foundry
This section of the Microsoft AI-102: Designing and Implementing a Microsoft Azure AI Solution exam covers building and deploying generative AI solutions using Azure AI Foundry. Below are study notes for each sub-topic, with links to Microsoft documentation, exam tips, and key facts.
Plan and Prepare for a Generative AI Solution
π Docs: Plan a generative AI solution
Overview
- Assess business needs and map them to Azure AI Foundry capabilities
- Decide between prebuilt models (e.g., GPT, Claude, Mistral) and fine-tuned/custom models
- Plan for Responsible AI principles: fairness, reliability, safety, privacy, inclusiveness, transparency
Key Points
- Ensure data compliance with Azureβs region availability
- Define KPIs: latency, cost, accuracy
- Choose the right model size (smaller for cost efficiency, larger for complex reasoning)
Exam Tip
Expect scenario-based questions asking you to match a use case to the correct model type or service.
Deploy a Hub, Project, and Necessary Resources with Azure AI Foundry
π Docs: Azure AI Foundry hubs and projects
Overview
- Hub: Central workspace for managing multiple AI projects
- Project: Contains assets (data, models, prompt flows)
- Requires linked Azure AI Search, Azure Storage, and sometimes Azure OpenAI resources
Key Facts
- A hub can support multiple projects
- Projects inherit security and networking from the hub
- Role-based access control (RBAC) applies at both hub and project levels
Limits
- Regional availability: Azure OpenAI is not available in all regions.
- Resource quotas: per-subscription and per-region.
Deploy the Appropriate Generative AI Model for Your Use Case
π Docs: Model catalog
Overview
- Choose models from the model catalog (e.g., GPT-4, GPT-35-Turbo, Phi-3, Mistral)
- Match task to model:
- Text generation β GPT series
- Code completion β Codex-like models
- Small footprint β Phi-3-mini
Key Points
- Pay attention to token limits (e.g., GPT-4 Turbo supports 128K context length)
- Pricing is per 1,000 tokens (input + output)
Exam Tip
Know model max context size and capabilities. Microsoft often tests model suitability.
Implement a Prompt Flow Solution
π Docs: Prompt flow in Azure AI Foundry
Overview
- Prompt flow = orchestration tool to design, test, evaluate, and deploy prompts
- Supports chaining prompts, grounding, and evaluation
- Includes visual editor and SDK integration
Key Facts
- Flows can include LLM calls, Python code, and external API calls
- Enables tracking, versioning, and collaboration
Use Case
Customer service bot with multi-turn reasoning.
Implement a RAG Pattern by Grounding a Model in Your Data
π Docs: Grounding and RAG in Azure AI
Overview
- RAG (Retrieval Augmented Generation): Combines LLM with enterprise data sources
- Steps:
- Ingest and chunk documents
- Embed using Azure OpenAI Embeddings
- Store in Azure AI Search
- Retrieve relevant chunks and pass to LLM
Key Points
- Improves factual accuracy and reduces hallucinations
- Embedding models:
text-embedding-ada-002
,text-embedding-3-large
- Vector store = Azure AI Search (supports hybrid search)
Exam Tip
Watch for scenarios: When to use fine-tuning vs. RAG? β RAG is better for dynamic/factual grounding.
Evaluate Models and Flows
π Docs: Evaluate models and flows
Overview
- Evaluate prompts and flows based on quality, reliability, safety
- Use human feedback + automated metrics (e.g., BLEU, ROUGE, perplexity)
Key Points
- Evaluation harness in Foundry helps test against sample datasets
- Supports A/B testing for prompts
Metrics
- Groundedness: factual correctness.
- Coherence: logical flow.
- Relevance: matches user intent.
Integrate Your Project into an Application with Azure AI Foundry SDK
π Docs: Azure AI Foundry SDK
Overview
- Python SDK enables programmatic control of projects, models, and flows
- Common tasks:
- Deploying prompt flows
- Running evaluations
- Accessing endpoints
Example
from azure.ai.foundry import FoundryClient
client = FoundryClient()
response = client.run_prompt_flow(flow_name="myflow", inputs={"question": "What is Azure AI Foundry?"})
print(response.output)
Key Points
- SDK integrates with CI/CD pipelines
- Supports endpoint management for integration with apps
Utilize Prompt Templates in Your Generative AI Solution
π Docs: Prompt templates
Overview
- Templates standardize prompts with placeholders
- Example:
You are a helpful assistant. Answer the following: {{user_question}}
Key Facts
- Reduces prompt injection risks
- Improves consistency and reusability
- Supports variables, loops, and conditional logic
Exam Tip
Microsoft may test knowledge of prompt injection mitigation β templates are one safeguard.
Quickβfire revision sheet
- π Know the difference between fine-tuning and RAG
- π Memorize model context limits and token pricing basics
- π Understand how hubs/projects/resources are structured
- π Be prepared for scenario-based questions on Responsible AI
- π Familiarize with evaluation metrics and when to apply them