<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"><channel><title>AI Explained: New Modules</title><description>Free, open-source AI education; new modules as they are published.</description><link>https://hmohamedansari.com/</link><language>en-gb</language><item><title>What is an LLM?</title><link>https://hmohamedansari.com/learn/foundations/what-is-an-llm/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/foundations/what-is-an-llm/</guid><description>Large Language Models are stateless text-transformation functions: they take text in and return text out, with no memory between calls. Understanding this one fact shapes every architectural decision you&apos;ll make with AI.</description><pubDate>Mon, 23 Mar 2026 00:00:00 GMT</pubDate><category>foundations</category><category>stable</category></item><item><title>How Prompts Work</title><link>https://hmohamedansari.com/learn/foundations/how-prompts-work/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/foundations/how-prompts-work/</guid><description>A prompt is not a question: it&apos;s a structured program. Understanding its anatomy (system instruction, conversation history, user message) lets you communicate intent reliably and debug output failures systematically.</description><pubDate>Mon, 23 Mar 2026 00:00:00 GMT</pubDate><category>foundations</category><category>stable</category></item><item><title>Models and Model Selection</title><link>https://hmohamedansari.com/learn/foundations/models-and-model-selection/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/foundations/models-and-model-selection/</guid><description>Not every task needs the most capable model. Understanding the capability-cost-latency tradeoff lets you pick the right model for each job, and avoid paying frontier prices for work a smaller model handles just as well.</description><pubDate>Mon, 23 Mar 2026 00:00:00 GMT</pubDate><category>foundations</category><category>volatile</category></item><item><title>Hallucinations and Model Reliability</title><link>https://hmohamedansari.com/learn/foundations/hallucinations-and-reliability/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/foundations/hallucinations-and-reliability/</guid><description>LLMs generate plausible text, not verified truth. Understanding why models hallucinate, and how to architect around it, is the single most important reliability concern in production AI systems.</description><pubDate>Mon, 23 Mar 2026 00:00:00 GMT</pubDate><category>foundations</category><category>stable</category></item><item><title>Structured Output and Tool Use</title><link>https://hmohamedansari.com/learn/foundations/structured-output-and-tool-use/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/foundations/structured-output-and-tool-use/</guid><description>Getting reliable, machine-readable output from an LLM requires more than asking nicely. Structured output and tool use turn a text generator into a component your application can depend on.</description><pubDate>Mon, 23 Mar 2026 00:00:00 GMT</pubDate><category>foundations</category><category>stable</category></item><item><title>Context and Memory Management</title><link>https://hmohamedansari.com/learn/foundations/context-and-memory-management/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/foundations/context-and-memory-management/</guid><description>LLMs are stateless: they have no memory between calls. Every form of &apos;memory&apos; in an AI application is something your code explicitly puts into the context window. Understanding how to manage that window is the core engineering skill behind every reliable AI system.</description><pubDate>Mon, 23 Mar 2026 00:00:00 GMT</pubDate><category>foundations</category><category>stable</category></item><item><title>Evaluating LLM Systems</title><link>https://hmohamedansari.com/learn/foundations/evaluating-llm-systems/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/foundations/evaluating-llm-systems/</guid><description>LLM outputs are probabilistic and hard to unit-test. Building a systematic evaluation practice, before you ship, and continuously in production, is what separates AI features that stay reliable from ones that silently degrade.</description><pubDate>Mon, 23 Mar 2026 00:00:00 GMT</pubDate><category>foundations</category><category>stable</category></item><item><title>Safety and Guardrails</title><link>https://hmohamedansari.com/learn/foundations/safety-and-guardrails/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/foundations/safety-and-guardrails/</guid><description>Safety in AI systems is not a single feature: it is a layered architecture. Understanding what the model handles automatically, what you must build, and where the gaps are is essential before shipping anything user-facing.</description><pubDate>Mon, 23 Mar 2026 00:00:00 GMT</pubDate><category>foundations</category><category>emerging</category></item><item><title>Prototype to Production Checklist</title><link>https://hmohamedansari.com/learn/foundations/prototype-to-production/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/foundations/prototype-to-production/</guid><description>A prototype that works in a demo is not a production system. This capstone synthesises every Foundations concept into a practical checklist: the gaps teams consistently miss when shipping their first AI feature.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>foundations</category><category>stable</category></item><item><title>What is RAG and Why</title><link>https://hmohamedansari.com/learn/rag/what-is-rag/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/rag/what-is-rag/</guid><description>LLMs know a lot, but they don&apos;t know your data. Retrieval-Augmented Generation is the pattern that fixes this: not by training the model on your data, but by finding the relevant pieces at query time and handing them directly to the model.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>rag</category><category>stable</category></item><item><title>Embeddings and Vector Search</title><link>https://hmohamedansari.com/learn/rag/embeddings-and-vector-search/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/rag/embeddings-and-vector-search/</guid><description>Semantic search, finding text by meaning rather than keywords, is the engine inside most RAG systems. Understanding how embeddings work and how vector databases store and query them is the foundation you need to build reliable retrieval.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>rag</category><category>stable</category></item><item><title>Chunking and Indexing</title><link>https://hmohamedansari.com/learn/rag/chunking-and-indexing/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/rag/chunking-and-indexing/</guid><description>You can&apos;t embed a whole document: you split it into pieces first. How you split determines what you can retrieve. The wrong chunking strategy is one of the most common reasons RAG systems fail to find the right answer even when the information clearly exists.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>rag</category><category>stable</category></item><item><title>Retrieval Quality: Dense, Sparse, and Hybrid</title><link>https://hmohamedansari.com/learn/rag/retrieval-quality/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/rag/retrieval-quality/</guid><description>Semantic search is powerful but not always the best retrieval method. Keyword search finds exact matches that embeddings miss. Re-ranking re-scores candidates with a slower but more accurate model. Understanding when to use each, and how to combine them, is what separates reliable RAG from fragile RAG.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>rag</category><category>stable</category></item><item><title>Prompting for RAG</title><link>https://hmohamedansari.com/learn/rag/prompting-for-rag/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/rag/prompting-for-rag/</guid><description>Retrieved chunks are only as useful as the instructions you give the model for using them. The grounding instruction, context format, citation pattern, and no-answer path are what turn a retrieval result into a reliable, trustworthy answer.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>rag</category><category>stable</category></item><item><title>Evaluating RAG Systems</title><link>https://hmohamedansari.com/learn/rag/evaluating-rag/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/rag/evaluating-rag/</guid><description>A fluent, well-formatted answer based on the wrong chunk is a failure, but it reads like a success. RAG evaluation requires two independent measurement tracks: retrieval quality and generation quality. Conflating them hides the real failure mode.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>rag</category><category>stable</category></item><item><title>Advanced RAG Patterns</title><link>https://hmohamedansari.com/learn/rag/advanced-rag-patterns/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/rag/advanced-rag-patterns/</guid><description>Basic RAG fails when queries are vague, answers span multiple documents, or context evolves across a conversation. Four patterns, multi-query retrieval, HyDE, contextual retrieval, and small-to-big, each fix a specific retrieval failure mode. Know which failure you have before reaching for a pattern.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>rag</category><category>emerging</category></item><item><title>Production RAG Checklist</title><link>https://hmohamedansari.com/learn/rag/production-rag-checklist/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/rag/production-rag-checklist/</guid><description>A RAG prototype that works on your test documents is not a production system. This capstone synthesises the full RAG track into a checklist: the gaps that consistently cause RAG failures after launch, and the order to address them.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>rag</category><category>stable</category></item><item><title>Protocol Landscape and MCP</title><link>https://hmohamedansari.com/learn/protocols/protocol-landscape/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/protocols/protocol-landscape/</guid><description>Before a model can call a tool, both sides have to agree on the contract: what the tool is called, what arguments it accepts, and what it returns. This module maps the protocol landscape and shows where MCP fits.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>protocols</category><category>volatile</category></item><item><title>Tool Schema Design</title><link>https://hmohamedansari.com/learn/protocols/tool-schema-design/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/protocols/tool-schema-design/</guid><description>The schema is not documentation: it is the instruction the model reads to decide whether to call your tool and what to pass. A bad schema causes wrong tool selections, invalid arguments, and hallucinated parameter values. This module covers what separates a production schema from a prototype one.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>protocols</category><category>stable</category></item><item><title>Tool Execution Patterns</title><link>https://hmohamedansari.com/learn/protocols/tool-execution-patterns/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/protocols/tool-execution-patterns/</guid><description>A single tool call is easy. Production tool use involves chains of calls, parallel execution, shared state, and the ever-present risk of runaway loops. This module covers the patterns that make multi-step tool execution reliable.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>protocols</category><category>stable</category></item><item><title>Real API Integration</title><link>https://hmohamedansari.com/learn/protocols/real-api-integration/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/protocols/real-api-integration/</guid><description>Wrapping a real API as a tool means handling all the things the happy path ignores: auth token expiry, rate limits, flaky networks, non-idempotent operations, and paginated results. This module covers the mechanics of building tool integrations that survive production.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>protocols</category><category>stable</category></item><item><title>Streaming and Async Tool Workflows</title><link>https://hmohamedansari.com/learn/protocols/streaming-and-async/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/protocols/streaming-and-async/</guid><description>Streaming gives users tokens as they arrive instead of waiting for the full response. Async tools let long-running operations run in the background. Both change how you wire together models and tools, and both have sharp edges that aren&apos;t obvious until you&apos;re in production.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>protocols</category><category>volatile</category></item><item><title>Security Boundaries</title><link>https://hmohamedansari.com/learn/protocols/security-boundaries/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/protocols/security-boundaries/</guid><description>Tools give models real capabilities: which means tool-using systems inherit the security risks of real software plus some new ones specific to AI. Prompt injection, over-privileged tools, and undelimited external content are the three failure modes that show up first. This module covers the boundaries that need to exist.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>protocols</category><category>stable</category></item><item><title>Production Operations</title><link>https://hmohamedansari.com/learn/protocols/production-operations/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/protocols/production-operations/</guid><description>A tool-using system has more moving parts than a simple prompt-response loop, and more things that can go wrong. This module covers the observability, cost management, and resilience patterns that keep tool integrations reliable after launch.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>protocols</category><category>stable</category></item><item><title>Testing and Reliability</title><link>https://hmohamedansari.com/learn/protocols/testing-and-reliability/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/protocols/testing-and-reliability/</guid><description>Tool-using systems are hard to test because the interesting behavior emerges from the interaction between the model and the tools, not from either alone. This module covers the testing strategy that catches real failures: schema drift, unexpected model behavior, and integration regressions.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>protocols</category><category>stable</category></item><item><title>What is an Agent</title><link>https://hmohamedansari.com/learn/agents/what-is-an-agent/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/agents/what-is-an-agent/</guid><description>An agent is not a smarter chatbot: it is a different execution model. This module defines what makes something agentic, maps the spectrum from single call to autonomous agent, and gives you the decision matrix to know which approach fits your problem.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>agents</category><category>emerging</category></item><item><title>Memory and State</title><link>https://hmohamedansari.com/learn/agents/memory-and-state/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/agents/memory-and-state/</guid><description>Memory is what separates a stateless chatbot from an agent that can work across sessions and build on past experience. This module covers the four memory types, how to manage the lifecycle of each, and the anti-patterns that cause agents to accumulate stale, conflicting, or poisoned state.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>agents</category><category>emerging</category></item><item><title>Planning and Decomposition</title><link>https://hmohamedansari.com/learn/agents/planning-and-decomposition/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/agents/planning-and-decomposition/</guid><description>Complex tasks fail when handed to an agent as a single goal. Planning is the process of decomposing a goal into executable steps: deciding what to do, in what order, and when to revise the plan based on what actually happens.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>agents</category><category>emerging</category></item><item><title>Multi-Agent Patterns</title><link>https://hmohamedansari.com/learn/agents/multi-agent-patterns/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/agents/multi-agent-patterns/</guid><description>A single agent hits limits: context windows fill, specialisation is hard, and long tasks become fragile. Multi-agent architectures solve this by distributing work, but they introduce coordination costs, trust boundaries, and new failure modes. This module covers the patterns that work in production.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>agents</category><category>emerging</category></item><item><title>Agent Failure Modes</title><link>https://hmohamedansari.com/learn/agents/agent-failure-modes/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/agents/agent-failure-modes/</guid><description>Agents fail in ways that are qualitatively different from single API calls: errors compound, loops consume unbounded resources, and failures can be invisible until they cause damage. This module catalogues the failure modes and the structural mitigations for each.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>agents</category><category>stable</category></item><item><title>Human-in-the-Loop</title><link>https://hmohamedansari.com/learn/agents/human-in-the-loop/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/agents/human-in-the-loop/</guid><description>Human oversight is not a bolt-on safety feature: it is an architectural primitive that determines what an agent is permitted to do autonomously and what requires a human decision. This module covers the design of approval gates, interrupt points, confidence escalation, and audit trails that make human oversight practical at scale.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>agents</category><category>stable</category></item><item><title>Production Agent Systems</title><link>https://hmohamedansari.com/learn/agents/production-agent-systems/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/agents/production-agent-systems/</guid><description>An agent that works in a demo fails in production the first time it crashes mid-task, gets retried with a duplicate side effect, or loses its state to a process restart. This module covers the durability semantics that separate toy agents from production systems.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>agents</category><category>emerging</category></item><item><title>Agent Evaluation</title><link>https://hmohamedansari.com/learn/agents/agent-evaluation/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/agents/agent-evaluation/</guid><description>Evaluating an agent is fundamentally different from evaluating a model. The question is not just &apos;was the answer correct?&apos; but &apos;did the agent take the right path to get there, and would it hold up under different conditions?&apos; This module covers offline trajectory evaluation and online production monitoring: the two distinct disciplines that together keep agent quality measurable.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>agents</category><category>emerging</category></item><item><title>Hosting Options</title><link>https://hmohamedansari.com/learn/infrastructure/hosting-options/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/infrastructure/hosting-options/</guid><description>Choosing where to run your model determines your cost structure, latency floor, and operational burden: understanding the tradeoffs between API inference, self-hosted, and cloud-managed endpoints lets you pick the right option for each workload rather than defaulting to whatever is easiest to start.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>infrastructure</category><category>emerging</category></item><item><title>Quantization &amp; Compression</title><link>https://hmohamedansari.com/learn/infrastructure/quantization-and-compression/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/infrastructure/quantization-and-compression/</guid><description>Quantization reduces the memory and compute cost of running a model by storing its weights in lower precision: understanding the tradeoffs between FP16, INT8, and INT4 and the methods used to get there lets you serve larger models on smaller hardware without silently breaking quality.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>infrastructure</category><category>emerging</category></item><item><title>Inference Serving</title><link>https://hmohamedansari.com/learn/infrastructure/inference-serving/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/infrastructure/inference-serving/</guid><description>Inference servers are not just web servers that happen to call a model: they implement specific memory management and scheduling algorithms that determine whether your GPU serves 5 requests per second or 50; understanding KV cache, PagedAttention, and continuous batching separates the teams who can scale from the teams who can&apos;t.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>infrastructure</category><category>emerging</category></item><item><title>Batching &amp; Throughput</title><link>https://hmohamedansari.com/learn/infrastructure/batching-and-throughput/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/infrastructure/batching-and-throughput/</guid><description>Throughput and latency are in direct tension in LLM serving: understanding how batching works, why continuous batching is the production default, and how to separate throughput benchmarks from latency benchmarks prevents the common mistake of optimizing one while silently destroying the other.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>infrastructure</category><category>emerging</category></item><item><title>Latency Optimization</title><link>https://hmohamedansari.com/learn/infrastructure/latency-optimization/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/infrastructure/latency-optimization/</guid><description>LLM latency has three distinct components, TTFT, TBT, and E2E, and different use cases require optimizing different ones; knowing which techniques reduce which component, and when prompt caching defeats itself, prevents wasted effort and avoids the most common serving regressions.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>infrastructure</category><category>emerging</category></item><item><title>Hardware Selection</title><link>https://hmohamedansari.com/learn/infrastructure/hardware-selection/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/infrastructure/hardware-selection/</guid><description>Choosing the wrong GPU tier, or sizing VRAM based on model weights alone, is the most common hardware mistake in LLM deployment; knowing the VRAM math, the GPU tiers, and when to use multi-GPU parallelism lets you right-size hardware before you need it rather than after an OOM in production.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>infrastructure</category><category>emerging</category></item><item><title>Containerization &amp; Deployment</title><link>https://hmohamedansari.com/learn/infrastructure/containerization-and-deployment/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/infrastructure/containerization-and-deployment/</guid><description>Containerizing an LLM inference server is fundamentally different from containerizing a web service; GPU passthrough, multi-stage weight management, and slow pod startup require different patterns for health checks, rolling deployments, and Kubernetes configuration that most teams learn by breaking production first.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>infrastructure</category><category>emerging</category></item><item><title>Scaling &amp; Cost Management</title><link>https://hmohamedansari.com/learn/infrastructure/scaling-and-cost-management/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/infrastructure/scaling-and-cost-management/</guid><description>LLM serving costs accumulate differently from typical web services; GPU-hours are expensive, autoscaling on CPU metrics is wrong, and scale-to-zero creates cold-start latency that makes it unsuitable for interactive workloads; knowing the right signals to scale on and how to build the cost math keeps infrastructure expenses from becoming a surprise.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>infrastructure</category><category>emerging</category></item><item><title>What Makes LLM Evaluation Hard</title><link>https://hmohamedansari.com/learn/evaluation/llm-eval-fundamentals/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/evaluation/llm-eval-fundamentals/</guid><description>Learn why LLM eval is structurally different from traditional ML testing, what the three axes of eval design are, and how to build a mental model for the rest of the track.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>evaluation</category><category>emerging</category></item><item><title>Building an Eval Dataset</title><link>https://hmohamedansari.com/learn/evaluation/eval-dataset-construction/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/evaluation/eval-dataset-construction/</guid><description>Learn to treat eval datasets as engineering artifacts: how to seed them, label them, version them, and keep them representative of real production traffic.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>evaluation</category><category>emerging</category></item><item><title>Automated Evaluation Methods</title><link>https://hmohamedansari.com/learn/evaluation/automated-eval-methods/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/evaluation/automated-eval-methods/</guid><description>Master the spectrum of automated eval techniques, from exact match and string overlap through semantic similarity and LLM-as-judge, and learn which method to apply for which task.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>evaluation</category><category>emerging</category></item><item><title>Tracing &amp; Structured Logging</title><link>https://hmohamedansari.com/learn/evaluation/tracing-and-logging/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/evaluation/tracing-and-logging/</guid><description>Learn to instrument LLM systems with structured traces that make debugging and performance analysis practical: what to log, how to structure it, and how to avoid PII liability.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>evaluation</category><category>emerging</category></item><item><title>Cost Attribution &amp; Token Budgets</title><link>https://hmohamedansari.com/learn/evaluation/cost-attribution/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/evaluation/cost-attribution/</guid><description>Learn to track, attribute, and control LLM API costs before the invoice surprises you: per-request tagging, per-feature aggregation, token budget enforcement, and anomaly alerting.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>evaluation</category><category>emerging</category></item><item><title>CI/CD Eval Gates</title><link>https://hmohamedansari.com/learn/evaluation/cicd-eval-gates/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/evaluation/cicd-eval-gates/</guid><description>Learn to build automated eval gates that block deployments when prompt changes, model upgrades, or RAG index updates regress quality: before they reach users.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>evaluation</category><category>emerging</category></item><item><title>Production Monitoring &amp; Drift Detection</title><link>https://hmohamedansari.com/learn/evaluation/production-monitoring/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/evaluation/production-monitoring/</guid><description>Learn to detect quality regressions, distribution shifts, and cost anomalies in live LLM systems before users report them: using metrics, statistical process control, and a sample-and-judge pipeline.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>evaluation</category><category>emerging</category></item><item><title>Red-teaming &amp; Adversarial Evaluation</title><link>https://hmohamedansari.com/learn/evaluation/red-teaming/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/evaluation/red-teaming/</guid><description>Learn to systematically discover failure modes in LLM systems before attackers do: how to run a red-team session, categorize findings, and convert every confirmed vulnerability into a permanent regression test.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>evaluation</category><category>emerging</category></item><item><title>The AI Threat Landscape</title><link>https://hmohamedansari.com/learn/safety/ai-threat-landscape/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/safety/ai-threat-landscape/</guid><description>Every LLM application has a multi-layer attack surface: model, context, tools, memory, and outputs. Understanding what attackers want and what they can do is the prerequisite to building defences that actually hold. This module maps the threat landscape and establishes why defence in depth is not optional.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>safety</category><category>emerging</category></item><item><title>Prompt Injection</title><link>https://hmohamedansari.com/learn/safety/prompt-injection/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/safety/prompt-injection/</guid><description>Prompt injection is the most prevalent attack class in LLM applications. It takes two forms: direct injection from user input, and indirect injection through retrieved documents or tool results. Both exploit the same root cause: the model cannot distinguish instructions from data when they share the same channel.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>safety</category><category>emerging</category></item><item><title>Jailbreaking and Policy Bypass</title><link>https://hmohamedansari.com/learn/safety/jailbreaking-and-policy-bypass/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/safety/jailbreaking-and-policy-bypass/</guid><description>Jailbreaking is the attempt to get a model to produce output that its alignment training or system prompt prohibit. No defence is permanent: the arms race between jailbreak techniques and countermeasures is ongoing. This module covers the attack taxonomy and the multi-layer defences that reduce, but never eliminate, the risk.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>safety</category><category>emerging</category></item><item><title>Data Privacy and PII</title><link>https://hmohamedansari.com/learn/safety/data-privacy-and-pii/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/safety/data-privacy-and-pii/</guid><description>LLM systems create new PII leakage vectors that traditional data protection controls do not cover: model memorisation, cross-user context leakage, and RAG pipelines that pull in customer records without scrubbing. This module covers detection, scrubbing, retention, and the vendor agreements that govern what happens to your data.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>safety</category><category>emerging</category></item><item><title>Guardrails Architecture</title><link>https://hmohamedansari.com/learn/safety/guardrails-architecture/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/safety/guardrails-architecture/</guid><description>Guardrails are controls on inputs, outputs, or both: classifiers, validators, and policy checks that run independently of the model. Designing a guardrails architecture means choosing which controls to apply, how to layer them for coverage and performance, and how to calibrate them so false positives do not kill legitimate use.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>safety</category><category>emerging</category></item><item><title>Supply Chain Security</title><link>https://hmohamedansari.com/learn/safety/supply-chain-security/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/safety/supply-chain-security/</guid><description>The AI supply chain, base model, fine-tuning data, adapters, Python packages, and API keys, has more attack surfaces than teams typically consider. A .pkl file is executable code. An unverified model weight can contain backdoors. This module covers the controls that keep your AI system trustworthy from training data to production inference.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>safety</category><category>emerging</category></item><item><title>Regulatory Landscape</title><link>https://hmohamedansari.com/learn/safety/regulatory-landscape/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/safety/regulatory-landscape/</guid><description>The regulatory environment for AI is moving quickly. The EU AI Act introduced risk tiers and mandatory requirements. GDPR has always applied to automated decision-making. The US has the NIST AI RMF. This module maps the landscape for a B2B SaaS product using LLMs: what you likely need to document, what you need to avoid, and where you need legal counsel.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>safety</category><category>emerging</category></item><item><title>Incident Response for AI Systems</title><link>https://hmohamedansari.com/learn/safety/incident-response/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/safety/incident-response/</guid><description>An AI incident is not a software incident: it involves model misbehaviour, safety violations, or data leakage, each with distinct root causes and remediation paths. This module covers detection, containment, investigation, and post-mortem structure for AI-specific incidents, and the one logging investment that makes all of it possible.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>safety</category><category>emerging</category></item><item><title>AI ROI: What Actually Gets Measured</title><link>https://hmohamedansari.com/learn/strategy/ai-roi/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/strategy/ai-roi/</guid><description>Most AI pilots show impressive returns that evaporate at scale. Understanding why, and how to measure value correctly, is the difference between AI investments that compound and ones that quietly fail.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>strategy</category><category>emerging</category></item><item><title>Buy vs Build vs Fine-tune</title><link>https://hmohamedansari.com/learn/strategy/buy-vs-build-vs-finetune/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/strategy/buy-vs-build-vs-finetune/</guid><description>Every AI capability involves a make-or-buy decision, but the options are more nuanced than they look. This module gives you a decision framework and total cost of ownership model for each path.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>strategy</category><category>emerging</category></item><item><title>Where AI Creates Durable Advantage</title><link>https://hmohamedansari.com/learn/strategy/ai-competitive-advantage/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/strategy/ai-competitive-advantage/</guid><description>Most AI features can be replicated by any competitor with API access. Durable advantage comes from the layer underneath: proprietary data, deep workflow integration, and feedback loops that compound over time.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>strategy</category><category>emerging</category></item><item><title>Team Structure and AI Capability</title><link>https://hmohamedansari.com/learn/strategy/team-structure-and-ai-capability/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/strategy/team-structure-and-ai-capability/</guid><description>How you organise your AI function determines what it can ship. This module maps the tradeoffs between centralised and federated models, defines the roles that actually matter, and gives you a maturity test for assessing whether your AI team can reach production.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>strategy</category><category>emerging</category></item><item><title>Managing AI Risk at the Org Level</title><link>https://hmohamedansari.com/learn/strategy/managing-ai-risk/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/strategy/managing-ai-risk/</guid><description>AI systems introduce risk categories that traditional software governance does not cover. This module maps the five risk categories, explains how to set risk appetite, and distinguishes real risk management from risk theatre.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>strategy</category><category>emerging</category></item><item><title>Communicating AI to Stakeholders</title><link>https://hmohamedansari.com/learn/strategy/communicating-ai-to-stakeholders/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/strategy/communicating-ai-to-stakeholders/</guid><description>The gap between what engineers know about AI systems and what stakeholders need to hear is where AI projects lose trust. This module gives you the frameworks to communicate outcomes, risk, cost, and failures in language that lands.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>strategy</category><category>emerging</category></item><item><title>AI Procurement and Vendor Evaluation</title><link>https://hmohamedansari.com/learn/strategy/ai-procurement-and-vendor-evaluation/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/strategy/ai-procurement-and-vendor-evaluation/</guid><description>Choosing an AI vendor on benchmark performance alone is one of the most reliable ways to end up with the wrong vendor. This module gives you a complete evaluation framework covering quality, pricing, data handling, SLAs, and exit planning.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>strategy</category><category>emerging</category></item><item><title>Building an AI-Ready Data Foundation</title><link>https://hmohamedansari.com/learn/strategy/ai-ready-data-foundation/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/strategy/ai-ready-data-foundation/</guid><description>Most AI ambitions stall not on model capability but on data readiness. This module gives you a practical checklist to assess whether your data is ready for AI, and explains why data infrastructure investment returns more than model investment for most organisations.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>strategy</category><category>emerging</category></item><item><title>How Vision-Language Models Work</title><link>https://hmohamedansari.com/learn/multimodal/vision-language-models/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/multimodal/vision-language-models/</guid><description>A vision-language model (VLM) combines a visual encoder with a language model: images are converted to token-like embeddings and fed directly into the same context window as text. Understanding this architecture explains why images cost more tokens than they appear to, and why resolution and tiling choices matter in production.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>multimodal</category><category>volatile</category></item><item><title>Working with Images in Production</title><link>https://hmohamedansari.com/learn/multimodal/images-in-production/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/multimodal/images-in-production/</guid><description>Sending an image to a VLM is trivial; building a production image pipeline that handles validation, preprocessing, output parsing, and failure modes is not. This module covers the full ingestion pipeline from receipt to parsed output, with emphasis on the silent failure modes that catch teams by surprise.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>multimodal</category><category>volatile</category></item><item><title>Audio and Speech AI</title><link>https://hmohamedansari.com/learn/multimodal/audio-and-speech-ai/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/multimodal/audio-and-speech-ai/</guid><description>The audio AI stack spans automatic speech recognition (ASR), text-to-speech (TTS), and the orchestration layer that connects them to language models. This module covers the key components, their production metrics, and the voice AI pipeline pattern that powers real-time conversational applications.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>multimodal</category><category>volatile</category></item><item><title>Multimodal Agents</title><link>https://hmohamedansari.com/learn/multimodal/multimodal-agents/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/multimodal/multimodal-agents/</guid><description>Multimodal agents extend the standard agent loop with perception across images and audio, and with actions that produce visual or spoken output. This module covers GUI agents, vision as a tool call, multimodal memory, and the specific failure modes that multimodal perception introduces into agent systems.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>multimodal</category><category>volatile</category></item><item><title>Multimodal Safety</title><link>https://hmohamedansari.com/learn/multimodal/multimodal-safety/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/multimodal/multimodal-safety/</guid><description>Images and audio introduce attack surfaces that text-only safety systems do not cover: injected instructions inside images, adversarial visual inputs, deepfakes, and PII embedded in non-text modalities. This module covers the threat model for multimodal inputs and the defensive patterns that close the gaps.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>multimodal</category><category>volatile</category></item><item><title>Multimodal Evaluation</title><link>https://hmohamedansari.com/learn/multimodal/multimodal-evaluation/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/multimodal/multimodal-evaluation/</guid><description>Evaluating multimodal AI is harder than evaluating text: there is no ground truth for &apos;describe this image&apos;, visual hallucinations are invisible without the source image, and labelling image datasets is expensive. This module covers evaluation approaches by task type, reference datasets, hallucination detection, and how to build a practical multimodal eval pipeline.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>multimodal</category><category>volatile</category></item><item><title>Serving Multimodal Models</title><link>https://hmohamedansari.com/learn/multimodal/serving-multimodal-models/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/multimodal/serving-multimodal-models/</guid><description>Serving a vision-language model is not the same as serving a text-only LLM: the vision encoder adds VRAM, image preprocessing adds latency, and variable image sizes complicate batching. This module covers the serving stack for VLMs and audio models, including the VRAM estimation mistakes that cause production OOMs.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>multimodal</category><category>volatile</category></item><item><title>The Multimodal Frontier</title><link>https://hmohamedansari.com/learn/multimodal/multimodal-frontier/</link><guid isPermaLink="true">https://hmohamedansari.com/learn/multimodal/multimodal-frontier/</guid><description>Multimodal AI is advancing faster than any other part of the field: native multimodality, video understanding, and real-time audio-visual interaction are moving from research to production on a timescale of months. This module covers where the field is heading and, more importantly, what durable knowledge to invest in when specific capabilities become outdated within a year.</description><pubDate>Tue, 24 Mar 2026 00:00:00 GMT</pubDate><category>multimodal</category><category>volatile</category></item></channel></rss>