跳到主要內容

學習、聯繫、建置

Microsoft Reactor

加入 Microsoft Reactor 並與開發人員即時互動

準備好開始使用 AI 和最新技術嗎? Microsoft Reactor 提供活動、訓練和社群資源,協助開發人員、企業家和初創公司建置 AI 技術等等。 加入我們!

學習、聯繫、建置

Microsoft Reactor

加入 Microsoft Reactor 並與開發人員即時互動

準備好開始使用 AI 和最新技術嗎? Microsoft Reactor 提供活動、訓練和社群資源,協助開發人員、企業家和初創公司建置 AI 技術等等。 加入我們!

返回

Python + Agents: Monitoring and evaluating agents

26 2月, 2026 | 6:30 下午 - 7:30 下午 (UTC) 國際標準時間

  • 格式:
  • alt##Livestream線上直播

主題: Agents

語言: 英文

In the third session of our Python + Agents series, we’ll focus on two essential components of building reliable agents: observability and evaluation.

We’ll begin with observability, using OpenTelemetry to capture traces, metrics, and logs from agent actions. You'll learn how to instrument your agents and use a local Aspire dashboard to identify slowdowns and failures.

From there, we’ll explore how to evaluate agent behavior using the Azure AI Evaluation SDK. You’ll see how to define evaluation criteria, run automated assessments over a set of tasks, and analyze the results to measure accuracy, helpfulness, and task success.

By the end of the session, you’ll have practical tools and workflows for monitoring, measuring, and improving your agents—so they’re not just functional, but dependable and verifiably effective.

必要條件

To follow along with the live examples, sign up for a free GitHub account. If you are brand new to generative AI with Python, start with our our 9-part Python + AI series, which covers LLMs, embedding models, RAG, tool calling, MCP, and more.

演講者

相關活動

下列活動也可能對您感興趣。 務必造訪我們的 Reactor 首頁 以查看所有可用的活動。

本頁面的一部分可能是機器翻譯或人工智能翻譯的.