Follow these steps to get started using these resources:
- Fork the Repository: Click
- Clone the Repository:
git clone https://github.com/microsoft/edgeai-for-beginners.git
- Join The Azure AI Foundry Discord and meet experts and fellow developers
Arabic | Bengali | Bulgarian | Burmese (Myanmar) | Chinese (Simplified) | Chinese (Traditional, Hong Kong) | Chinese (Traditional, Macau) | Chinese (Traditional, Taiwan) | Croatian | Czech | Danish | Dutch | Finnish | French | German | Greek | Hebrew | Hindi | Hungarian | Indonesian | Italian | Japanese | Korean | Malay | Marathi | Nepali | Norwegian | Persian (Farsi) | Polish | Portuguese (Brazil) | Portuguese (Portugal) | Punjabi (Gurmukhi) | Romanian | Russian | Serbian (Cyrillic) | Slovak | Slovenian | Spanish | Swahili | Swedish | Tagalog (Filipino) | Thai | Turkish | Ukrainian | Urdu | Vietnamese
If you wish to have additional translations languages supported are listed here
Welcome to EdgeAI for Beginners β your comprehensive journey into the transformative world of Edge Artificial Intelligence. This course bridges the gap between powerful AI capabilities and practical, real-world deployment on edge devices, empowering you to harness AI's potential directly where data is generated and decisions need to be made.
This course takes you from fundamental concepts to production-ready implementations, covering:
- Small Language Models (SLMs) optimized for edge deployment
- Hardware-aware optimization across diverse platforms
- Real-time inference with privacy-preserving capabilities
- Production deployment strategies for enterprise applications
Edge AI represents a paradigm shift that addresses critical modern challenges:
- Privacy & Security: Process sensitive data locally without cloud exposure
- Real-time Performance: Eliminate network latency for time-critical applications
- Cost Efficiency: Reduce bandwidth and cloud computing expenses
- Resilient Operations: Maintain functionality during network outages
- Regulatory Compliance: Meet data sovereignty requirements
Edge AI refers to running AI algorithms and language models locally on hardware, close to where data is generated without relying on cloud resources for inference. It reduces latency, enhances privacy, and enables real-time decision-making.
- On-device inference: AI models run on edge devices (phones, routers, microcontrollers, industrial PCs)
- Offline capability: Functions without persistent internet connectivity
- Low latency: Immediate responses suited for real-time systems
- Data sovereignty: Keeps sensitive data local, improving security and compliance
SLMs like Phi-4, Mistral-7B, and Gemma are optimized versions of larger LLMsβtrained or distilled for:
- Reduced memory footprint: Efficient use of limited edge device memory
- Lower compute demand: Optimized for CPU and edge GPU performance
- Faster startup times: Quick initialization for responsive applications
They unlock powerful NLP capabilities while meeting the constraints of:
- Embedded systems: IoT devices and industrial controllers
- Mobile devices: Smartphones and tablets with offline capabilities
- IoT Devices: Sensors and smart devices with limited resources
- Edge servers: Local processing units with limited GPU resources
- Personal Computers: Desktop and laptop deployment scenarios
Module | Topic | Focus Area | Key Content | Level | Duration |
---|---|---|---|---|---|
π 00 | Introduction to EdgeAI | Foundation & Context | EdgeAI Overview β’ Industry Applications β’ SLM Introduction β’ Learning Objectives | Beginner | 1-2 hrs |
π 01 | EdgeAI Fundamentals | Cloud vs Edge AI comparison | EdgeAI Fundamentals β’ Real World Case Studies β’ Implementation Guide β’ Edge Deployment | Beginner | 3-4 hrs |
π§ 02 | SLM Model Foundations | Model families & architecture | Phi Family β’ Qwen Family β’ Gemma Family β’ BitNET β’ ΞΌModel β’ Phi-Silica | Beginner | 4-5 hrs |
π 03 | SLM Deployment Practice | Local & cloud deployment | Advanced Learning β’ Local Environment β’ Cloud Deployment | Intermediate | 4-5 hrs |
βοΈ 04 | Model Optimization Toolkit | Cross-platform optimization | Introduction β’ Llama.cpp β’ Microsoft Olive β’ OpenVINO β’ Apple MLX β’ Workflow Synthesis | Intermediate | 5-6 hrs |
π§ 05 | SLMOps Production | Production operations | SLMOps Introduction β’ Model Distillation β’ Fine-tuning β’ Production Deployment | Advanced | 5-6 hrs |
π€ 06 | AI Agents & Function Calling | Agent frameworks & MCP | Agent Introduction β’ Function Calling β’ Model Context Protocol | Advanced | 4-5 hrs |
π» 07 | Platform Implementation | Cross-platform samples | AI Toolkit β’ Foundry Local β’ Windows Development | Advanced | 3-4 hrs |
π 08 | Foundry Local Toolkit | Production-ready samples | Sample applications (see details below) | Expert | 8-10 hrs |
- 01: REST Chat Quickstart
- 02: OpenAI SDK Integration
- 03: Model Discovery & Benchmarking
- 04: Chainlit RAG Application
- 05: Multi-Agent Orchestration
- 06: Models-as-Tools Router
- 07: Direct API Client
- 08: Windows 11 Chat App
- 09: Advanced Multi-Agent System
- 10: Foundry Tools Framework
Comprehensive hands-on workshop materials with production-ready implementations:
- Workshop Guide - Complete learning objectives, outcomes, and resource navigation
- Python Samples (6 sessions) - Updated with best practices, error handling, and comprehensive documentation
- Jupyter Notebooks (8 interactive) - Step-by-step tutorials with benchmarks and performance monitoring
- Session Guides - Detailed markdown guides for each workshop session
- Validation Tools - Scripts to verify code quality and run smoke tests
What You'll Build:
- Local AI chat applications with streaming support
- RAG pipelines with quality evaluation (RAGAS)
- Multi-model benchmarking and comparison tools
- Multi-agent orchestration systems
- Intelligent model routing with task-based selection
- Total Duration: 36-45 hours
- Beginner Path: Modules 01-02 (7-9 hours)
- Intermediate Path: Modules 03-04 (9-11 hours)
- Advanced Path: Modules 05-07 (12-15 hours)
- Expert Path: Module 08 (8-10 hours)
- Edge AI Architecture: Design local-first AI systems with cloud integration
- Model Optimization: Quantize and compress models for edge deployment (85% speed boost, 75% size reduction)
- Multi-Platform Deployment: Windows, mobile, embedded, and cloud-edge hybrid systems
- Production Operations: Monitoring, scaling, and maintaining edge AI in production
- Foundry Local Chat Apps: Windows 11 native application with model switching
- Multi-Agent Systems: Coordinator with specialist agents for complex workflows
- RAG Applications: Local document processing with vector search
- Model Routers: Intelligent selection between models based on task analysis
- API Frameworks: Production-ready clients with streaming and health monitoring
- Cross-Platform Tools: LangChain/Semantic Kernel integration patterns
Manufacturing β’ Healthcare β’ Autonomous Vehicles β’ Smart Cities β’ Mobile Apps
Recommended Learning Path (20-30 hours total):
- π Introduction (Introduction.md): EdgeAI foundation + industry context + learning framework
- π Foundation (Modules 01-02): EdgeAI concepts + SLM model families
- βοΈ Optimization (Modules 03-04): Deployment + quantization frameworks
- π Production (Modules 05-06): SLMOps + AI agents + function calling
- π» Implementation (Modules 07-08): Platform samples + Foundry Local toolkit
Each module includes theory, hands-on exercises, and production-ready code samples.
Technical Roles: EdgeAI Solutions Architect β’ ML Engineer (Edge) β’ IoT AI Developer β’ Mobile AI Developer
Industry Sectors: Manufacturing 4.0 β’ Healthcare Tech β’ Autonomous Systems β’ FinTech β’ Consumer Electronics
Portfolio Projects: Multi-agent systems β’ Production RAG apps β’ Cross-platform deployment β’ Performance optimization
edgeai-for-beginners/
βββ π introduction.md # Foundation: EdgeAI Overview & Learning Framework
βββ π Module01-04/ # Fundamentals β SLMs β Deployment β Optimization
βββ π§ Module05-06/ # SLMOps β AI Agents β Function Calling
βββ π» Module07/ # Platform Samples (VS Code, Windows, Jetson, Mobile)
βββ π Module08/ # Foundry Local Toolkit + 10 Comprehensive Samples
β βββ samples/01-06/ # Foundation: REST, SDK, RAG, Agents, Routing
β βββ samples/07-10/ # Advanced: API Client, Windows App, Enterprise Agents, Tools
βββ π translations/ # Multi-language support (8+ languages)
βββ π STUDY_GUIDE.md # Structured learning paths & time allocation
β
Progressive Learning: Theory β Practice β Production deployment
β
Real Case Studies: Microsoft, Japan Airlines, enterprise implementations
β
Hands-on Samples: 50+ examples, 10 comprehensive Foundry Local demos
β
Performance Focus: 85% speed improvements, 75% size reductions
β
Multi-Platform: Windows, mobile, embedded, cloud-edge hybrid
β
Production Ready: Monitoring, scaling, security, compliance frameworks
π Study Guide Available: Structured 20-hour learning path with time allocation guidance and self-assessment tools.
EdgeAI represents the future of AI deployment: local-first, privacy-preserving, and efficient. Master these skills to build the next generation of intelligent applications.
Our team produces other courses! Check out:
- MCP for Beginners
- AI Agents For Beginners
- Generative AI for Beginners using .NET
- Generative AI for Beginners using JavaScript
- Generative AI for Beginners
- ML for Beginners
- Data Science for Beginners
- AI for Beginners
- Cybersecurity for Beginners
- Web Dev for Beginners
- IoT for Beginners
- XR Development for Beginners
- Mastering GitHub Copilot for AI Paired Programming
- Mastering GitHub Copilot for C#/.NET Developers
- Choose Your Own Copilot Adventure
If you get stuck or have any questions about building AI apps, join:
If you have product feedback or errors while building visit: