AI AgentsΒΆ
Build intelligent agents that can reason, plan, and use tools to accomplish complex tasks
π― Learning ObjectivesΒΆ
By the end of this phase, you will:
β Understand what AI agents are and how they differ from simple chatbots
β Design and implement tool/function schemas for agents
β Build agents that can use multiple tools to accomplish tasks
β Implement the ReAct (Reasoning + Acting) pattern
β Create multi-step agents with memory and state management
β Use agent frameworks (LangChain, LangGraph, OpenAI Agents SDK, CrewAI)
β Understand MCP and modern agent interoperability patterns
β Compare managed vs self-hosted agent stacks
β Understand the 2026 agentic platform landscape (OpenHands, OpenCode, Lingxi, mini-swe-agent, computer-use agents)
β Evaluate and observe agents in production
β Deploy production-ready AI agents
π What Are AI Agents?ΒΆ
AI Agent: An autonomous system that can:
Perceive its environment (receive inputs)
Reason about what action to take
Act by using tools/functions
Learn from feedback to improve
Chatbot vs AgentΒΆ
Feature |
Chatbot |
AI Agent |
|---|---|---|
Capability |
Responds to queries |
Takes actions |
Tools |
None |
Can use external tools |
Autonomy |
Passive |
Proactive |
Memory |
Conversation history |
Persistent state + context |
Reasoning |
Single-turn |
Multi-step planning |
Example |
βTell me about Parisβ |
βBook me a flight to Parisβ |
π Phase ContentΒΆ
NotebooksΒΆ
-
What are AI agents?
Agent architecture
Simple agent example
Agent design patterns
-
Tool schema design
OpenAI Function Calling API
Tool selection strategies
Error handling
-
ReAct: Reasoning + Acting
Chain-of-thought with tools
Multi-step problem solving
Building a research agent
-
LangChain agents
LangGraph for workflows
Managed vs self-hosted agent APIs
Open-source framework comparison
Interop protocols (MCP, A2A)
-
Coordinating multiple agents
Agent communication protocols
Task delegation
Building an agent team
06_mcp_model_context_protocol.ipynb
MCP: the emerging standard for AI tool integration (2026)
Connecting LLMs to external tools and data sources
Building MCP servers and clients
07_openai_agents_sdk_langgraph.ipynb
OpenAI Agents SDK
LangGraph stateful agents
Comparing agent frameworks
-
Reasoning-capable models for agent workflows
Deliberate planning vs fast-response models
When longer thinking improves tool use
09_autonomous_agents_2026.ipynb
State of the art: autonomous agents in 2026
OpenHands, OpenCode, Lingxi, mini-swe-agent, and computer-use agents
Production patterns and best practices
Future directions
-
Why agent evaluation is hard (non-determinism, side-effects)
Four dimensions: task success, trajectory quality, tool correctness, safety
Offline eval with LLM-as-Judge scoring
Online eval: observability, tracing, cost tracking
Frameworks: promptfoo, LangSmith, Braintrust, Arize Phoenix
Assignments & PracticeΒΆ
assignment.md - Build a production-ready AI agent (6 bonus options incl. MCP & eval)
challenges.md - 9 hands-on challenges (ββ to βββββ)
QuizzesΒΆ
Pre-Quiz: Assess baseline knowledge
Post-Quiz: Validate learning outcomes
π οΈ Tools Youβll UseΒΆ
OpenAI Function Calling - Native tool use
LangChain - Agent framework
LangGraph - Workflow orchestration
OpenAI Agents SDK - Lightweight multi-agent handoffs and tracing
CrewAI - Role-based multi-agent coordination
Google ADK / Semantic Kernel - Additional framework families to be aware of
MCP - Standard tool connectivity across agent runtimes
OpenHands / OpenCode / mini-swe-agent / Lingxi - Important 2026 agentic coding platforms
2026 Agent Topics To KnowΒΆ
Managed agent APIs vs self-hosted frameworks
Open-source agentic coding platforms vs proprietary IDE agents
MCP for tool integration and A2A for agent delegation
Agent observability: tracing, tool-call inspection, latency and cost tracking
Agent evaluation: task success, trajectory quality, tool correctness, and safety gates
Long-running and proactive agents rather than single-request assistants
π Real-World ApplicationsΒΆ
1. Customer Support AgentΒΆ
Answer FAQs
Query knowledge base
Create support tickets
Escalate to humans
2. Research AssistantΒΆ
Search web/papers
Summarize findings
Generate reports
Cite sources
3. Code Generation AgentΒΆ
Understand requirements
Write code
Run tests
Debug errors
Deploy to production
4. Data Analysis AgentΒΆ
Load datasets
Exploratory analysis
Generate visualizations
Statistical testing
Create reports
5. Personal AssistantΒΆ
Check calendar
Send emails
Book meetings
Set reminders
Research topics
π Project: Build Your Own AgentΒΆ
Youβll build one of these agents:
SQL Agent - Natural language β SQL queries β Results β Insights
Research Agent - Topic β Search β Summarize β Report
Coding Agent - Requirements β Code β Test β Fix β Deploy
Customer Service Agent - Query β Knowledge base β Response β Ticket
β±οΈ Time CommitmentΒΆ
Videos/Reading: 4 hours
Notebooks: 6 hours
Assignment: 8 hours
Challenges: 4-12 hours (optional)
Total: ~18-30 hours
π PrerequisitesΒΆ
Before starting this phase, ensure you understand:
β LLM Basics (Phase 11: Prompt Engineering)
β API Usage (Python, REST APIs)
β JSON (Tool schemas are JSON)
β Async Programming (For concurrent tool calls)
β RAG Systems (Phase 8) - helpful but not required
π ResourcesΒΆ
DocumentationΒΆ
PapersΒΆ
VideosΒΆ
CommunityΒΆ
π Learning PathΒΆ
Week 1: Fundamentals
βββ Day 1-2: What are agents? (Notebook 1)
βββ Day 3-4: Function calling (Notebook 2)
βββ Day 5-7: ReAct pattern (Notebook 3)
Week 2: Frameworks & Protocols
βββ Day 1-2: Agent frameworks & no-code builders (Notebook 4)
βββ Day 3-4: Multi-agent systems & A2A (Notebook 5)
βββ Day 5: MCP deep-dive (Notebook 6)
βββ Day 6-7: OpenAI Agents SDK & LangGraph (Notebook 7)
Week 3: Advanced & Evaluation
βββ Day 1-2: Reasoning models in agent loops (Notebook 8)
βββ Day 3-4: Autonomous agents 2026 (Notebook 9)
βββ Day 5: Agent evaluation & safety (Notebook 10)
βββ Day 6-7: Assignment β build, evaluate & deploy
β AssessmentΒΆ
Pre-Quiz (10 questions)ΒΆ
Test your baseline knowledge of:
Agent concepts
Tool use patterns
API design
Post-Quiz (10 questions)ΒΆ
Validate your mastery of:
Agent architecture
Function calling implementation
ReAct pattern
Production best practices
Assignment (100 points)ΒΆ
Build a production-ready agent with:
Multiple tools (30 pts)
Error handling (20 pts)
Memory management (20 pts)
Evaluation & testing (30 pts)
π Success CriteriaΒΆ
By the end of this phase, you should be able to:
β Explain the difference between chatbots and agents
β Design effective tool schemas
β Implement function calling with OpenAI API
β Build a ReAct agent from scratch
β Use LangChain/LangGraph for complex workflows
β Implement agent memory and state
β Evaluate agents with LLM-as-Judge and trajectory scoring
β Deploy a production agent
β Debug common agent issues
π Next StepsΒΆ
After completing this phase:
Phase 16: Model Evaluation & Metrics
Phase 17: Debugging AI Systems
Phase 18: Low-Code AI Tools (Gradio, Streamlit)
Or explore advanced topics:
Multi-agent collaboration
Agent fine-tuning
Reinforcement learning for agents
Human-in-the-loop systems
Ready to build intelligent agents? Letβs go! π€β¨