AI Foundations

Welcome to AI Foundations Course

This comprehensive course explores the key concepts needed to build production-grade AI applications. Through a combination of theoretical foundations and practical applications, you'll build the skills necessary to understand and create AI-powered solutions.

Course Objectives

Establish a Shared Foundation
Ensure all team membersβ€”scientists, data engineers, and software engineersβ€”have a strong, common understanding of modern AI/LLM concepts, terminology, and best practices.
Why? This enables more effective collaboration, clearer communication, and faster consensus when designing, reviewing, or iterating on AI architectures.
Enable Business-Facing AI Application Development
Equip the team with the knowledge and practical skills needed to design, build, and deploy AI-powered applications that directly address business needs.
Why? This bridges the gap between technical capability and business value, ensuring our solutions are relevant and impactful.
Improve Project Scoping and Communication
Develop the ability to accurately estimate timelines, resource needs, and technical risks for AI projects. Empower team members to communicate requirements, dependencies, and trade-offs clearly to product managers and stakeholders.
Why? This leads to more predictable delivery, better alignment with business priorities, and fewer surprises during execution.
Accelerate Development with Modern AI Coding Tools
Integrate and adopt AI-powered coding tools (e.g., AmazonQ, Cline, Cursor etc.) into daily workflows to boost productivity and code quality.
Why? Leveraging these tools allows us to focus on higher-level design and problem-solving, while reducing manual effort and boilerplate.
Foster a Culture of Experimentation and Continuous Learning
Encourage team members to experiment with new prompting techniques, architectures, and evaluation methods. Share learnings and best practices across the team to continuously raise the bar for AI application quality and innovation.
Why? The AI field is evolving rapidly; a culture of curiosity and sharing ensures we stay ahead and adapt quickly.

Flavors of AI

AI is not a single technology β€” it spans a spectrum of capabilities and purposes. Understanding these flavors helps clarify what today's systems can and cannot do, and where the field is heading.

Flavors of AI Wheel
Flavor Question it answers Examples
🟠 Predictive What will happen? Fraud detection, demand forecasting, recommendations
πŸ”΅ Prescriptive What should we do? Dynamic pricing, supply chain optimization, treatment plans
🟑 Causal Why did it happen? A/B test analysis, root cause analysis, scientific discovery
🟣 Generative β˜… What can be created? Claude, ChatGPT, DALLΒ·E, Amazon Nova, Sora
πŸ”΄ Agentic β˜… What actions should I take? AI coworkers, autonomous research agents, software agents

β˜… This course focuses on Generative and Agentic AI β€” the two flavors driving today's wave of AI application development.

The Evolution of Artificial Intelligence

AI has evolved through the convergence of algorithmic breakthroughs and advances in computing infrastructure. From John McCarthy's 1956 Dartmouth workshop to today's agentic systems β€” here's how we got here.

1950s-1960s:

The Birth of AI

🧠 Algorithms: John McCarthy coins the term "Artificial Intelligence" at the 1956 Dartmouth Conference, establishing AI as a formal discipline. Early rule-based systems and symbolic reasoning programs follow.
πŸ’Ύ Infrastructure: First transistor-based computers like IBM 704 (1954) and LISP machines enable early AI research.
1990s-2000s:

Machine Learning Era

🧠 Algorithms: Rise of statistical ML approaches, Support Vector Machines (SVM), and early neural networks enable learning from data.
πŸ–₯️ Infrastructure: Personal computers become powerful enough for ML tasks. Early GPU computing (NVIDIA CUDA, 2006) sets stage for deep learning revolution.
2012:

AlexNet & the Deep Learning Moment

🧠 Algorithms: AlexNet wins ImageNet by a massive margin, proving deep convolutional neural networks could outperform decades of hand-crafted computer vision. The deep learning era begins in earnest.
πŸ—οΈ Infrastructure: NVIDIA GPUs (CUDA) make training large neural networks practical. Rise of hyperscale cloud providers (AWS, Azure, GCP) democratizes compute access.
2017:

"Attention Is All You Need" β€” The Transformer

πŸ€– Algorithms: Google researchers publish the Transformer architecture, replacing recurrent networks with self-attention. This single paper directly enables BERT, GPT, and every modern LLM that followed.
πŸ’» Infrastructure: Distributed training frameworks (PyTorch, TensorFlow) and TPUs/A100 GPUs enable training models at previously impossible scale.
2022:

The ChatGPT Moment β€” GenAI Goes Mainstream

πŸ€– Algorithms: ChatGPT launches and reaches 100M users in 2 months β€” the fastest product adoption in history. RLHF (Reinforcement Learning from Human Feedback) makes LLMs genuinely useful and safe to deploy broadly.
🌐 Infrastructure: Cloud providers launch managed AI services (Amazon Bedrock, Azure OpenAI, Vertex AI), making frontier models accessible via API without owning any infrastructure.
2023-2025:

Multimodal AI & Cloud Platforms

πŸ€– Algorithms: Mixture of Experts (MoE) architecture (GPT-4, Gemini, Mixtral) enables frontier-level capability at a fraction of the compute cost β€” only the most relevant "expert" sub-networks activate per token. Then in January 2025, DeepSeek R1 lands as the open-source shock: a Chinese lab matches GPT-4 performance using MoE, open-sources the weights, and triggers a 17% single-day drop in NVIDIA's stock. Proof that frontier AI is no longer a closed club.
🌐 Infrastructure: Hyperscale cloud providers make frontier models API-accessible without owning any hardware. Amazon Bedrock (unified model access + Knowledge Bases + Guardrails), Azure OpenAI, and Google Vertex AI become the default deployment layer for enterprise AI applications.
2025-Present:

Agentic AI & The Coding Inflection Point

πŸ€– Algorithms: AI shifts from tool to colleague. Claude Sonnet 3.7 (Feb 2025) is the first model to genuinely reason through complex codebases end-to-end β€” not autocomplete, but architectural understanding. Claude Code takes this further: an agentic coding system that navigates multi-file projects, runs tests, fixes its own errors, and executes on intent. For software developers, this is as significant as the ChatGPT moment was for general users. Broadly, agentic frameworks (observe β†’ reason β†’ act loops) enable AI to take autonomous multi-step actions across tools, APIs, and systems.
🌐 Infrastructure: Cloud providers launch dedicated agentic building blocks β€” Amazon Bedrock AgentCore (Runtime, Memory, Gateway, Observability), enabling teams to deploy, orchestrate, and monitor AI agents in production at scale.

Course Structure

This course is designed specifically for scientists, software and data engineers. Through a carefully structured learning path, you'll gain both theoretical knowledge and practical skills needed to build production-grade AI applications.

Part 1: Foundational Concepts

Master the core concepts of modern AI development. Starting from fundamentals, you'll progressively build knowledge of language models, prompt engineering, AI agents, embeddings & RAG, and MCP β€” the essential building blocks for creating AI applications on AWS.

Key Outcomes:

  • Understand how LLMs work under the hood
  • Master effective prompt engineering
  • Build and evaluate agentic LLM applications
  • Connect LLMs to your data with embeddings and RAG
  • Understand the what, why, and how of MCP

Introduction to Large Language Models

Learn about the architecture, capabilities, and limitations of Large Language Models. Understand the fundamental concepts behind these powerful AI systems that are driving innovation across industries.

Start Module

Prompt Engineering Guide

Master the art of effectively communicating with AI models through carefully crafted prompts. Learn strategies and techniques to get the most accurate and useful responses from language models.

Start Module

Agentic LLM Applications

Discover how agentic LLM applications extend language models with memory, tool use, and iterative decision cycles to autonomously solve complex, multi-step tasks. This module covers how agentic LLM applications go beyond single-step or workflow-based applications to deliver adaptable, goal-driven AI solutions.

Start Module

Embeddings & Retrieval-Augmented Generation (RAG)

Understand how embeddings encode semantic meaning, how vector stores enable similarity search, and how to build RAG pipelines that give LLMs access to your private data β€” without fine-tuning. Covers Amazon Titan Embeddings, OpenSearch, and AWS Bedrock Knowledge Bases.

Start Module

Developer's Guide to Model Context Protocol (MCP)

Learn how to use MCP to build robust, maintainable AI applications. Understand the core principles of MCP and how it enables standardized communication between AI models and tools.

Start Module

Part 2: Building AI Applications

Part 1 gave you the concepts. Part 2 is where you build. Each module introduces a real tool or practice, and the part culminates in a capstone project β€” a working chat agent for this course, built on AWS using everything you've learned.

Key Outcomes:

  • Build agents using Amazon Bedrock, AgentCore, and Strands SDK
  • Use Spec-Driven Development with Kiro to go from spec to working code
  • Monitor, evaluate, and guard AI applications in production
  • Ship a real RAG-powered chat agent using AWS-native services

AWS AI Building Blocks

Get hands-on with the AWS services that power production AI applications. Covers Amazon Bedrock (foundation model API, model selection), Bedrock Agents (action groups, knowledge base integration), AgentCore Runtime and Gateway, and the Strands SDK for building agentic applications on AWS.

Spec-Driven Development with Kiro

Learn to build AI applications the right way β€” starting from a spec. Kiro is Amazon's AI-powered developer tool that supports both vibe coding and structured spec-driven development. This module covers the Kiro CLI and IDE, how to write effective specs, and how spec-driven workflows lead to better, more maintainable AI-powered code.

LLMOps: Monitoring, Evaluation & Guardrails

Understand the operational side of AI applications. Learn how to trace and monitor agent behavior with CloudWatch, evaluate output quality, manage prompt versions with Bedrock Prompt Management, and enforce safety boundaries with Bedrock Guardrails. What you learn here applies directly to the capstone project.

Capstone: Build the Course Chat Agent

Put it all together. Crawl and index this course's content into a Bedrock Knowledge Base, build a Strands SDK agent grounded in that knowledge, expose it through AgentCore Gateway as an MCP endpoint, instrument it with LLMOps tooling, and write the whole thing spec-first using Kiro. The result: a working AI assistant that knows this course, deployed on AWS.