<?xml version="1.0" encoding="UTF-8"?><?xml-stylesheet href="https://feeds.captivate.fm/style.xsl" type="text/xsl"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:sy="http://purl.org/rss/1.0/modules/syndication/" xmlns:podcast="https://podcastindex.org/namespace/1.0"><channel><atom:link href="https://feeds.captivate.fm/memriq-inference-brief-engineering/" rel="self" type="application/rss+xml"/><title><![CDATA[The Memriq AI Inference Brief – Engineering Edition]]></title><podcast:guid>73ca278e-2ff2-5930-a961-67e4715fee75</podcast:guid><lastBuildDate>Wed, 18 Feb 2026 16:23:57 +0000</lastBuildDate><generator>Captivate.fm</generator><language><![CDATA[en]]></language><copyright><![CDATA[Copyright 2025 Memriq AI]]></copyright><managingEditor>Keith Bourne</managingEditor><itunes:summary><![CDATA[The Memriq AI Inference Brief – Engineering Edition is a weekly deep dive into the technical guts of modern AI systems: retrieval-augmented generation (RAG), vector databases, knowledge graphs, agents, memory systems, and more. A rotating panel of AI engineers and data scientists breaks down architectures, frameworks, and patterns from real-world projects so you can ship more intelligent systems, faster.]]></itunes:summary><image><url>https://artwork.captivate.fm/04981885-e7b7-4eb7-af52-e0d393698735/memriq-inference-brief-engineering.jpg</url><title>The Memriq AI Inference Brief – Engineering Edition</title><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm]]></link></image><itunes:image href="https://artwork.captivate.fm/04981885-e7b7-4eb7-af52-e0d393698735/memriq-inference-brief-engineering.jpg"/><itunes:owner><itunes:name>Keith Bourne</itunes:name></itunes:owner><itunes:author>Keith Bourne</itunes:author><description>The Memriq AI Inference Brief – Engineering Edition is a weekly deep dive into the technical guts of modern AI systems: retrieval-augmented generation (RAG), vector databases, knowledge graphs, agents, memory systems, and more. A rotating panel of AI engineers and data scientists breaks down architectures, frameworks, and patterns from real-world projects so you can ship more intelligent systems, faster.</description><link>https://memriq-inference-brief-engineering.captivate.fm</link><atom:link href="https://pubsubhubbub.appspot.com" rel="hub"/><itunes:subtitle><![CDATA[RAG pipelines, agent memory, knowledge graphs — the technical details that matter. Let's dig in.]]></itunes:subtitle><itunes:explicit>false</itunes:explicit><itunes:type>episodic</itunes:type><itunes:category text="Technology"></itunes:category><itunes:category text="News"><itunes:category text="Tech News"/></itunes:category><itunes:category text="Education"><itunes:category text="Self-Improvement"/></itunes:category><podcast:locked>no</podcast:locked><podcast:medium>podcast</podcast:medium><item><title>Kaizen at Digital Speed: Engineering the Agentic Enterprise Operating System</title><itunes:title>Kaizen at Digital Speed: Engineering the Agentic Enterprise Operating System</itunes:title><description><![CDATA[<p>In this episode of Memriq Inference Digest — Engineering Edition, we dive into the transformational role of engineers in the age of the agentic enterprise. Discover how continuous improvement at digital speed reshapes engineering from shipping code to building self-improving workflows powered by autonomous AI agents.</p><p><strong>In this episode:</strong></p><p>- Explore the shift from feature delivery to workflow orchestration in agentic systems</p><p>- Understand the five technical pillars every agent engineer must master</p><p>- Learn why operational literacy and governance are critical skills for engineers</p><p>- Contrast 'tool-first' versus 'operating-system-first' engineering approaches</p><p>- Get practical steps to prepare yourself for the future of agent-driven enterprises</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>- Autonomous AI agents</p><p>- Workflow orchestration and architecture</p><p>- Observability frameworks (logging, metrics, traces)</p><p>- Evaluation and continuous testing harnesses</p><p>- Governance models and policy gates</p><p><strong>Timestamps:</strong></p><p>0:00 Introduction &amp; episode overview</p><p>2:30 Why agentification matters now</p><p>5:15 The evolving role of engineers in the agentic enterprise</p><p>8:45 The five technical pillars: workflow, integration, observability, evaluation, governance</p><p>14:30 Engineering paths: tool-first vs operating-system-first</p><p>17:00 Practical preparation roadmap for engineers</p><p>19:30 Closing thoughts &amp; next steps</p><p><strong>Resources:</strong></p><p><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e11-agentic-enterprise-e" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a> by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e11-agentic-enterprise-e" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>This podcast is brought to you by <a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e11-agentic-enterprise-e" rel="noopener noreferrer" target="_blank">Memriq.ai</a> - AI consultancy and content studio building tools and resources for AI practitioners.</p>]]></description><content:encoded><![CDATA[<p>In this episode of Memriq Inference Digest — Engineering Edition, we dive into the transformational role of engineers in the age of the agentic enterprise. Discover how continuous improvement at digital speed reshapes engineering from shipping code to building self-improving workflows powered by autonomous AI agents.</p><p><strong>In this episode:</strong></p><p>- Explore the shift from feature delivery to workflow orchestration in agentic systems</p><p>- Understand the five technical pillars every agent engineer must master</p><p>- Learn why operational literacy and governance are critical skills for engineers</p><p>- Contrast 'tool-first' versus 'operating-system-first' engineering approaches</p><p>- Get practical steps to prepare yourself for the future of agent-driven enterprises</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>- Autonomous AI agents</p><p>- Workflow orchestration and architecture</p><p>- Observability frameworks (logging, metrics, traces)</p><p>- Evaluation and continuous testing harnesses</p><p>- Governance models and policy gates</p><p><strong>Timestamps:</strong></p><p>0:00 Introduction &amp; episode overview</p><p>2:30 Why agentification matters now</p><p>5:15 The evolving role of engineers in the agentic enterprise</p><p>8:45 The five technical pillars: workflow, integration, observability, evaluation, governance</p><p>14:30 Engineering paths: tool-first vs operating-system-first</p><p>17:00 Practical preparation roadmap for engineers</p><p>19:30 Closing thoughts &amp; next steps</p><p><strong>Resources:</strong></p><p><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e11-agentic-enterprise-e" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a> by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e11-agentic-enterprise-e" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>This podcast is brought to you by <a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e11-agentic-enterprise-e" rel="noopener noreferrer" target="_blank">Memriq.ai</a> - AI consultancy and content studio building tools and resources for AI practitioners.</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/agentic-enterprise-engineering]]></link><guid isPermaLink="false">7b7281dd-41cb-464c-ab2b-dc9a4f54a17c</guid><itunes:image href="https://artwork.captivate.fm/3518940c-96e1-48b3-9183-040ee8ed254e/artwork-20260218-104647.jpg"/><pubDate>Mon, 16 Feb 2026 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/7b7281dd-41cb-464c-ab2b-dc9a4f54a17c.mp3" length="23393324" type="audio/mpeg"/><itunes:duration>19:30</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:season>1</itunes:season><itunes:episode>11</itunes:episode><podcast:episode>11</podcast:episode><podcast:season>1</podcast:season><podcast:transcript url="https://transcripts.captivate.fm/transcript/2228e694-57f8-42e6-967c-1fedb89f0d58/index.html" type="text/html"/></item><item><title>Opus 4.6 Deep Dive: Memory, Reasoning &amp; Multi-Agent AI Architectures</title><itunes:title>Opus 4.6 Deep Dive: Memory, Reasoning &amp; Multi-Agent AI Architectures</itunes:title><description><![CDATA[<p>Unlock the potential of Anthropic's Claude Opus 4.6, a breakthrough AI model designed for deep reasoning and multi-agent orchestration with a massive one million token context window. Discover how this update transforms agent stack design by introducing adaptive effort tuning, advanced memory management, and role discipline in multi-model pipelines.</p><p><strong>In this episode:</strong></p><p>- Explore Opus 4.6’s unique ‘effort’ parameter and its role in controlling deep reasoning workloads</p><p>- Understand how Opus 4.6 integrates large context windows and subagent orchestration for complex workflows</p><p>- Compare Opus 4.6 with OpenAI’s GPT-5.2 to weigh trade-offs in cost, multimodality, and reasoning depth</p><p>- Learn practical deployment strategies and model role assignments for efficient multi-agent pipelines</p><p>- Hear real-world success stories from enterprises leveraging Opus 4.6 in production</p><p>- Review open challenges like cost governance, migration complexity, and multi-agent safety</p><p>Key tools &amp; technologies mentioned: Anthropic Claude Opus 4.6, OpenAI GPT-5.2, GitHub Copilot, Retrieval-Augmented Generation, Adaptive Thinking, Effort Parameter, Multi-Agent AI Pipelines</p><p><strong>Timestamps:</strong></p><p>[00:00] Introduction &amp; Episode Overview</p><p>[02:30] The 'Effort' Parameter &amp; Overthinking Feature</p><p>[06:00] Why Opus 4.6 Matters Now: Long Context &amp; Reasoning Boost</p><p>[09:30] Architecting Multi-Model Agent Pipelines</p><p>[12:45] Head-to-Head: Opus 4.6 vs GPT-5.2</p><p>[15:00] Under the Hood: Technical Innovations</p><p>[17:30] Real-World Impact &amp; Use Cases</p><p>[19:45] Practical Tips &amp; Open Challenges</p><p><strong>Resources:</strong></p><p>- <u><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e10-opus-4-6-ai-architec" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a></u> by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e10-opus-4-6-ai-architec" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- This podcast is brought to you by <a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e10-opus-4-6-ai-architec" rel="noopener noreferrer" target="_blank">Memriq.ai</a> - AI consultancy and content studio building tools and resources for AI practitioners.</p>]]></description><content:encoded><![CDATA[<p>Unlock the potential of Anthropic's Claude Opus 4.6, a breakthrough AI model designed for deep reasoning and multi-agent orchestration with a massive one million token context window. Discover how this update transforms agent stack design by introducing adaptive effort tuning, advanced memory management, and role discipline in multi-model pipelines.</p><p><strong>In this episode:</strong></p><p>- Explore Opus 4.6’s unique ‘effort’ parameter and its role in controlling deep reasoning workloads</p><p>- Understand how Opus 4.6 integrates large context windows and subagent orchestration for complex workflows</p><p>- Compare Opus 4.6 with OpenAI’s GPT-5.2 to weigh trade-offs in cost, multimodality, and reasoning depth</p><p>- Learn practical deployment strategies and model role assignments for efficient multi-agent pipelines</p><p>- Hear real-world success stories from enterprises leveraging Opus 4.6 in production</p><p>- Review open challenges like cost governance, migration complexity, and multi-agent safety</p><p>Key tools &amp; technologies mentioned: Anthropic Claude Opus 4.6, OpenAI GPT-5.2, GitHub Copilot, Retrieval-Augmented Generation, Adaptive Thinking, Effort Parameter, Multi-Agent AI Pipelines</p><p><strong>Timestamps:</strong></p><p>[00:00] Introduction &amp; Episode Overview</p><p>[02:30] The 'Effort' Parameter &amp; Overthinking Feature</p><p>[06:00] Why Opus 4.6 Matters Now: Long Context &amp; Reasoning Boost</p><p>[09:30] Architecting Multi-Model Agent Pipelines</p><p>[12:45] Head-to-Head: Opus 4.6 vs GPT-5.2</p><p>[15:00] Under the Hood: Technical Innovations</p><p>[17:30] Real-World Impact &amp; Use Cases</p><p>[19:45] Practical Tips &amp; Open Challenges</p><p><strong>Resources:</strong></p><p>- <u><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e10-opus-4-6-ai-architec" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a></u> by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e10-opus-4-6-ai-architec" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- This podcast is brought to you by <a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e10-opus-4-6-ai-architec" rel="noopener noreferrer" target="_blank">Memriq.ai</a> - AI consultancy and content studio building tools and resources for AI practitioners.</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/opus-4-6-ai-architecture]]></link><guid isPermaLink="false">3718bd02-f499-4254-bdab-10fe66bbffe2</guid><itunes:image href="https://artwork.captivate.fm/4341dca7-33b0-40ae-8b38-f1d9596f29ea/artwork-20260207-140440.jpg"/><pubDate>Mon, 09 Feb 2026 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/3718bd02-f499-4254-bdab-10fe66bbffe2.mp3" length="24216044" type="audio/mpeg"/><itunes:duration>20:11</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:season>1</itunes:season><itunes:episode>10</itunes:episode><podcast:episode>10</podcast:episode><podcast:season>1</podcast:season><podcast:transcript url="https://transcripts.captivate.fm/transcript/9a72a435-a267-4163-9349-0556a86e2ee2/index.html" type="text/html"/></item><item><title>Moltbook Uncovered: Lessons from the AI Social Network Experiment</title><itunes:title>Moltbook Uncovered: Lessons from the AI Social Network Experiment</itunes:title><description><![CDATA[<p>Explore Moltbook, the groundbreaking AI social network where autonomous agents debate, self-organize, and evolve their own culture — revealing critical insights for developers building agentic systems. In this episode, we unpack Moltbook’s architecture, emergent behaviors, and the leadership challenges posed by autonomous AI social dynamics.</p><p><strong>In this episode:</strong></p><p>- What makes Moltbook a unique multi-agent AI social network and why it matters now</p><p>- The technical core: personality templates, interaction graphs, and reinforcement learning</p><p>- Trade-offs between emergent social AI and traditional rule-based multi-agent systems</p><p>- Real-world applications and the cost, governance, and risk considerations for leaders</p><p>- Practical strategies and tooling advice for developers experimenting with agentic AI</p><p>- Open challenges including unpredictability, bias, and evaluation in emergent AI cultures</p><p><strong>Key tools &amp; technologies: </strong>Transformer-based large language models, multi-agent reinforcement learning frameworks, interaction graph data structures</p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction to Moltbook and agentic AI social networks</p><p>03:30 - The AI social drama and emergent behaviors in Moltbook</p><p>08:15 - Technical deep dive: architecture and agent design</p><p>12:00 - Payoff metrics and emergent cultures</p><p>14:30 - Leadership reality checks and governance implications</p><p>17:00 - Practical applications and tech battle scenario</p><p>19:30 - Open problems and final insights</p><p><strong>Resources:</strong></p><p>- <a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e9-moltbook-ai-social-n" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a> by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e9-moltbook-ai-social-n" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- This podcast is brought to you by <a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e9-moltbook-ai-social-n" rel="noopener noreferrer" target="_blank">Memriq.ai</a> - AI consultancy and content studio building tools and resources for AI practitioners.</p>]]></description><content:encoded><![CDATA[<p>Explore Moltbook, the groundbreaking AI social network where autonomous agents debate, self-organize, and evolve their own culture — revealing critical insights for developers building agentic systems. In this episode, we unpack Moltbook’s architecture, emergent behaviors, and the leadership challenges posed by autonomous AI social dynamics.</p><p><strong>In this episode:</strong></p><p>- What makes Moltbook a unique multi-agent AI social network and why it matters now</p><p>- The technical core: personality templates, interaction graphs, and reinforcement learning</p><p>- Trade-offs between emergent social AI and traditional rule-based multi-agent systems</p><p>- Real-world applications and the cost, governance, and risk considerations for leaders</p><p>- Practical strategies and tooling advice for developers experimenting with agentic AI</p><p>- Open challenges including unpredictability, bias, and evaluation in emergent AI cultures</p><p><strong>Key tools &amp; technologies: </strong>Transformer-based large language models, multi-agent reinforcement learning frameworks, interaction graph data structures</p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction to Moltbook and agentic AI social networks</p><p>03:30 - The AI social drama and emergent behaviors in Moltbook</p><p>08:15 - Technical deep dive: architecture and agent design</p><p>12:00 - Payoff metrics and emergent cultures</p><p>14:30 - Leadership reality checks and governance implications</p><p>17:00 - Practical applications and tech battle scenario</p><p>19:30 - Open problems and final insights</p><p><strong>Resources:</strong></p><p>- <a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e9-moltbook-ai-social-n" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a> by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e9-moltbook-ai-social-n" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- This podcast is brought to you by <a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e9-moltbook-ai-social-n" rel="noopener noreferrer" target="_blank">Memriq.ai</a> - AI consultancy and content studio building tools and resources for AI practitioners.</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/moltbook-ai-social-network]]></link><guid isPermaLink="false">5f2a69c2-8ea6-4a6b-bf65-46c7bf492356</guid><itunes:image href="https://artwork.captivate.fm/96570352-329c-411c-992c-47814c3744bd/artwork-20260206-193816.jpg"/><pubDate>Mon, 02 Feb 2026 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/5f2a69c2-8ea6-4a6b-bf65-46c7bf492356.mp3" length="31829324" type="audio/mpeg"/><itunes:duration>26:31</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:season>1</itunes:season><itunes:episode>9</itunes:episode><podcast:episode>9</podcast:episode><podcast:season>1</podcast:season><podcast:transcript url="https://transcripts.captivate.fm/transcript/8f9f5fa4-08b7-42c2-96ae-c1be06095251/index.html" type="text/html"/></item><item><title>Agent-Driven UI Testing: What Changes &amp; Which Stacks Are Ready?</title><itunes:title>Agent-Driven UI Testing: What Changes &amp; Which Stacks Are Ready?</itunes:title><description><![CDATA[<p>UI testing has long been a pain point for engineering teams—expensive to write, brittle, and hard to maintain. In this episode of Memriq Inference Digest - Edition, we explore how AI-powered agents are transforming end-to-end (E2E) click-through testing by automating test planning, generation, and repair, making UI testing more scalable and sustainable. We also compare how different technology stacks like React/Next.js and Flutter support these new agent-driven approaches.</p><p><strong>In this episode, we cover:</strong></p><p>- Why traditional E2E UI tests often fail to catch real user issues despite existing for years</p><p>- How Playwright’s Planner, Generator, and Healer agents automate test lifecycle and maintenance</p><p>- The impact of UI framework choices on agent-driven testing success, especially React/Next.js vs Flutter</p><p>- Practical trade-offs between AI code-generation tools and runtime UI interaction agents</p><p>- Real-world examples and engineering best practices to make UI tests robust and maintainable</p><p>- Open challenges and the future direction of agent-driven UI testing</p><p><strong>Key tools and technologies discussed:</strong></p><p>- Playwright v1.56 AI agents (Planner, Generator, Healer)</p><p>- React and Next.js web frameworks</p><p>- Flutter’s flutter_test and integration_test frameworks</p><p>- Vitest, Jest, MSW for test runners and mocking</p><p>- AI coding assistants like Claude Code and GitHub Copilot</p><p><br></p><p><strong>Timestamps:</strong></p><p>0:00 – Introduction to Agent-Driven UI Testing</p><p>3:30 – Why Traditional E2E Tests Often Fail</p><p>6:45 – Playwright’s Planner, Generator &amp; Healer Explained</p><p>10:15 – Framework Readiness: React/Next.js vs Flutter</p><p>13:00 – Comparing AI Code Gen and Agent-Driven Testing</p><p>15:30 – Real-World Use Cases and Engineering Insights</p><p>18:00 – Open Challenges &amp; The Future of Agent-Driven Testing</p><p>20:00 – Closing Thoughts and Book Recommendation</p><p><br></p><p><strong>Resources:</strong></p><p>- <a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e8-agent-driven-ui-test" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a> by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e8-agent-driven-ui-test" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- This podcast is brought to you by <a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e8-agent-driven-ui-test" rel="noopener noreferrer" target="_blank">Memriq.ai</a> - AI consultancy and content studio building tools and resources for AI practitioners.</p>]]></description><content:encoded><![CDATA[<p>UI testing has long been a pain point for engineering teams—expensive to write, brittle, and hard to maintain. In this episode of Memriq Inference Digest - Edition, we explore how AI-powered agents are transforming end-to-end (E2E) click-through testing by automating test planning, generation, and repair, making UI testing more scalable and sustainable. We also compare how different technology stacks like React/Next.js and Flutter support these new agent-driven approaches.</p><p><strong>In this episode, we cover:</strong></p><p>- Why traditional E2E UI tests often fail to catch real user issues despite existing for years</p><p>- How Playwright’s Planner, Generator, and Healer agents automate test lifecycle and maintenance</p><p>- The impact of UI framework choices on agent-driven testing success, especially React/Next.js vs Flutter</p><p>- Practical trade-offs between AI code-generation tools and runtime UI interaction agents</p><p>- Real-world examples and engineering best practices to make UI tests robust and maintainable</p><p>- Open challenges and the future direction of agent-driven UI testing</p><p><strong>Key tools and technologies discussed:</strong></p><p>- Playwright v1.56 AI agents (Planner, Generator, Healer)</p><p>- React and Next.js web frameworks</p><p>- Flutter’s flutter_test and integration_test frameworks</p><p>- Vitest, Jest, MSW for test runners and mocking</p><p>- AI coding assistants like Claude Code and GitHub Copilot</p><p><br></p><p><strong>Timestamps:</strong></p><p>0:00 – Introduction to Agent-Driven UI Testing</p><p>3:30 – Why Traditional E2E Tests Often Fail</p><p>6:45 – Playwright’s Planner, Generator &amp; Healer Explained</p><p>10:15 – Framework Readiness: React/Next.js vs Flutter</p><p>13:00 – Comparing AI Code Gen and Agent-Driven Testing</p><p>15:30 – Real-World Use Cases and Engineering Insights</p><p>18:00 – Open Challenges &amp; The Future of Agent-Driven Testing</p><p>20:00 – Closing Thoughts and Book Recommendation</p><p><br></p><p><strong>Resources:</strong></p><p>- <a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e8-agent-driven-ui-test" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a> by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e8-agent-driven-ui-test" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- This podcast is brought to you by <a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e8-agent-driven-ui-test" rel="noopener noreferrer" target="_blank">Memriq.ai</a> - AI consultancy and content studio building tools and resources for AI practitioners.</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/agent-driven-ui-testing]]></link><guid isPermaLink="false">53e1f0a8-21a2-4aa9-8b2a-db128b36be38</guid><itunes:image href="https://artwork.captivate.fm/0473009a-93a3-4f4a-89fb-b54fcc375ff6/artwork-20260207-121423.jpg"/><pubDate>Mon, 26 Jan 2026 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/53e1f0a8-21a2-4aa9-8b2a-db128b36be38.mp3" length="25196204" type="audio/mpeg"/><itunes:duration>21:00</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:season>1</itunes:season><itunes:episode>8</itunes:episode><podcast:episode>8</podcast:episode><podcast:season>1</podcast:season><podcast:transcript url="https://transcripts.captivate.fm/transcript/72b7b8b3-d576-497d-8914-65b903570d5f/index.html" type="text/html"/></item><item><title>Belief States Uncovered: Internal Knowledge &amp; Uncertainty in AI Agents</title><itunes:title>Belief States Uncovered: Internal Knowledge &amp; Uncertainty in AI Agents</itunes:title><description><![CDATA[<p>Uncertainty is not just noise—it's the internal state that guides AI decision-making. In this episode of Memriq Inference Digest, we explore belief states, a foundational concept that enables AI systems to represent and reason about incomplete information effectively. From classical Bayesian filtering to cutting-edge neural planners like BetaZero, we unpack how belief states empower intelligent agents in real-world, uncertain environments.</p><p><strong>In this episode:</strong></p><p>- Understand the core concept of belief states and their role in AI under partial observability</p><p>- Compare symbolic, probabilistic, and neural belief state representations and their trade-offs</p><p>- Dive into practical implementations including Bayesian filtering, particle filters, and neural implicit beliefs</p><p>- Explore integrating belief states with CoALA memory systems for conversational AI</p><p>- Discuss real-world applications in robotics, autonomous vehicles, and dialogue systems</p><p>- Highlight open challenges and research frontiers including scalability, calibration, and multi-agent belief reasoning</p><p><strong>Key tools/technologies mentioned:</strong></p><p>- Partially Observable Markov Decision Processes (POMDPs)</p><p>- Bayesian filtering methods: Kalman filters, particle filters</p><p>- Neural networks: RNNs, Transformers</p><p>- Generative models: VAEs, GANs, diffusion models</p><p>- BetaZero and Monte Carlo tree search</p><p>- AGM belief revision framework</p><p>- I-POMDPs for multi-agent settings</p><p>- CoALA agentic memory architecture</p><p><strong>Resources:</strong></p><ol><li data-list="bullet"><span class="ql-ui" contenteditable="false"></span><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e7-belief-states-ai-unc" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a> by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e7-belief-states-ai-unc" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</li><li data-list="bullet"><span class="ql-ui" contenteditable="false"></span>This podcast is brought to you by <a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e7-belief-states-ai-unc" rel="noopener noreferrer" target="_blank">Memriq.ai</a> - AI consultancy and content studio building tools and resources for AI practitioners.</li></ol><br/>]]></description><content:encoded><![CDATA[<p>Uncertainty is not just noise—it's the internal state that guides AI decision-making. In this episode of Memriq Inference Digest, we explore belief states, a foundational concept that enables AI systems to represent and reason about incomplete information effectively. From classical Bayesian filtering to cutting-edge neural planners like BetaZero, we unpack how belief states empower intelligent agents in real-world, uncertain environments.</p><p><strong>In this episode:</strong></p><p>- Understand the core concept of belief states and their role in AI under partial observability</p><p>- Compare symbolic, probabilistic, and neural belief state representations and their trade-offs</p><p>- Dive into practical implementations including Bayesian filtering, particle filters, and neural implicit beliefs</p><p>- Explore integrating belief states with CoALA memory systems for conversational AI</p><p>- Discuss real-world applications in robotics, autonomous vehicles, and dialogue systems</p><p>- Highlight open challenges and research frontiers including scalability, calibration, and multi-agent belief reasoning</p><p><strong>Key tools/technologies mentioned:</strong></p><p>- Partially Observable Markov Decision Processes (POMDPs)</p><p>- Bayesian filtering methods: Kalman filters, particle filters</p><p>- Neural networks: RNNs, Transformers</p><p>- Generative models: VAEs, GANs, diffusion models</p><p>- BetaZero and Monte Carlo tree search</p><p>- AGM belief revision framework</p><p>- I-POMDPs for multi-agent settings</p><p>- CoALA agentic memory architecture</p><p><strong>Resources:</strong></p><ol><li data-list="bullet"><span class="ql-ui" contenteditable="false"></span><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e7-belief-states-ai-unc" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a> by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e7-belief-states-ai-unc" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</li><li data-list="bullet"><span class="ql-ui" contenteditable="false"></span>This podcast is brought to you by <a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e7-belief-states-ai-unc" rel="noopener noreferrer" target="_blank">Memriq.ai</a> - AI consultancy and content studio building tools and resources for AI practitioners.</li></ol><br/>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/belief-states-ai-uncertainty]]></link><guid isPermaLink="false">87730423-a38d-4b16-bb27-0c4bd1bdbaf7</guid><itunes:image href="https://artwork.captivate.fm/fdbec216-a927-4221-9cac-6d369587664e/artwork-20260110-052049.jpg"/><pubDate>Mon, 19 Jan 2026 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/87730423-a38d-4b16-bb27-0c4bd1bdbaf7.mp3" length="44233484" type="audio/mpeg"/><itunes:duration>36:52</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:season>1</itunes:season><itunes:episode>7</itunes:episode><podcast:episode>7</podcast:episode><podcast:season>1</podcast:season><podcast:transcript url="https://transcripts.captivate.fm/transcript/5a034f03-e351-4080-b4c8-38663f4e1455/index.html" type="text/html"/></item><item><title>Recursive Language Models: A Paradigm Shift for Agentic AI Scalability</title><itunes:title>Recursive Language Models: A Paradigm Shift for Agentic AI Scalability</itunes:title><description><![CDATA[<p>Discover how Recursive Language Models (RLMs) are fundamentally changing the way AI systems handle ultra-long contexts and complex reasoning. In this episode, we unpack why RLMs enable models to programmatically query massive corpora—two orders of magnitude larger than traditional transformers—delivering higher accuracy and cost efficiency for agentic AI applications.</p><p>In this episode:</p><p>- Explore the core architectural shift behind RLMs and how they externalize context via sandboxed Python environments</p><p>- Compare RLMs against other long-context approaches like Gemini 1.5 Pro, Longformer, BigBird, and RAG</p><p>- Dive into technical trade-offs including latency, cost variability, and verification overhead</p><p>- Hear real-world use cases in legal discovery, codebase analysis, and research synthesis</p><p>- Get practical tips on tooling with RLM official repo, Modal and Prime sandboxes, and hybrid workflows</p><p>- Discuss open challenges and future research directions for optimizing RLM deployments</p><p>Key tools and technologies mentioned:</p><p>- Recursive Language Model (RLM) official GitHub repo</p><p>- Modal and Prime sandboxed execution environments</p><p>- GPT-5 and GPT-5-mini models</p><p>- Gemini 1.5 Pro, Longformer, BigBird architectures</p><p>- Retrieval-Augmented Generation (RAG)</p><p>- Prime Intellect context folding</p><p>- MemGPT, LLMLingua token compression</p><p>Timestamps:</p><p>00:00 - Introduction to Recursive Language Models and agentic AI</p><p>03:15 - The paradigm shift: externalizing context and recursive querying</p><p>07:30 - Benchmarks and performance comparisons with other long-context models</p><p>11:00 - Under the hood: how RLMs orchestrate recursive sub-LLM calls</p><p>14:20 - Real-world applications: legal, code, and research use cases</p><p>16:45 - Technical trade-offs: latency, cost, and verification</p><p>18:30 - Toolbox and best practices for engineers</p><p>20:15 - Future directions and closing thoughts</p><p>Resources:</p><p><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e6-recursive-language-m" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a> by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e6-recursive-language-m" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>This podcast is brought to you by <a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e6-recursive-language-m" rel="noopener noreferrer" target="_blank">Memriq.ai</a> - AI consultancy and content studio building tools and resources for AI practitioners.</p><p>Stay tuned and keep pushing the boundaries of AI engineering with Memriq Inference Digest!</p>]]></description><content:encoded><![CDATA[<p>Discover how Recursive Language Models (RLMs) are fundamentally changing the way AI systems handle ultra-long contexts and complex reasoning. In this episode, we unpack why RLMs enable models to programmatically query massive corpora—two orders of magnitude larger than traditional transformers—delivering higher accuracy and cost efficiency for agentic AI applications.</p><p>In this episode:</p><p>- Explore the core architectural shift behind RLMs and how they externalize context via sandboxed Python environments</p><p>- Compare RLMs against other long-context approaches like Gemini 1.5 Pro, Longformer, BigBird, and RAG</p><p>- Dive into technical trade-offs including latency, cost variability, and verification overhead</p><p>- Hear real-world use cases in legal discovery, codebase analysis, and research synthesis</p><p>- Get practical tips on tooling with RLM official repo, Modal and Prime sandboxes, and hybrid workflows</p><p>- Discuss open challenges and future research directions for optimizing RLM deployments</p><p>Key tools and technologies mentioned:</p><p>- Recursive Language Model (RLM) official GitHub repo</p><p>- Modal and Prime sandboxed execution environments</p><p>- GPT-5 and GPT-5-mini models</p><p>- Gemini 1.5 Pro, Longformer, BigBird architectures</p><p>- Retrieval-Augmented Generation (RAG)</p><p>- Prime Intellect context folding</p><p>- MemGPT, LLMLingua token compression</p><p>Timestamps:</p><p>00:00 - Introduction to Recursive Language Models and agentic AI</p><p>03:15 - The paradigm shift: externalizing context and recursive querying</p><p>07:30 - Benchmarks and performance comparisons with other long-context models</p><p>11:00 - Under the hood: how RLMs orchestrate recursive sub-LLM calls</p><p>14:20 - Real-world applications: legal, code, and research use cases</p><p>16:45 - Technical trade-offs: latency, cost, and verification</p><p>18:30 - Toolbox and best practices for engineers</p><p>20:15 - Future directions and closing thoughts</p><p>Resources:</p><p><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e6-recursive-language-m" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a> by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e6-recursive-language-m" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>This podcast is brought to you by <a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e6-recursive-language-m" rel="noopener noreferrer" target="_blank">Memriq.ai</a> - AI consultancy and content studio building tools and resources for AI practitioners.</p><p>Stay tuned and keep pushing the boundaries of AI engineering with Memriq Inference Digest!</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/recursive-language-models-agentic-ai]]></link><guid isPermaLink="false">8f4b743f-b618-4e0c-a860-58699471caf3</guid><itunes:image href="https://artwork.captivate.fm/1bf610eb-32f3-4716-af0c-ffb8a5de6b9c/artwork-20260109-222846.jpg"/><pubDate>Mon, 12 Jan 2026 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/8f4b743f-b618-4e0c-a860-58699471caf3.mp3" length="24816044" type="audio/mpeg"/><itunes:duration>20:41</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:season>1</itunes:season><itunes:episode>6</itunes:episode><podcast:episode>6</podcast:episode><podcast:season>1</podcast:season><podcast:transcript url="https://transcripts.captivate.fm/transcript/306836f5-b1e5-4f7a-9705-8854284cbe62/index.html" type="text/html"/></item><item><title>Evaluating Agentic AI: DeepEval, RAGAS &amp; TruLens Frameworks Compared</title><itunes:title>Evaluating Agentic AI: DeepEval, RAGAS &amp; TruLens Frameworks Compared</itunes:title><description><![CDATA[<p><strong># Evaluating Agentic AI: DeepEval, RAGAS &amp; TruLens Frameworks Compared</strong></p><p>In this episode of Memriq Inference Digest - Engineering Edition, we explore the cutting-edge evaluation frameworks designed for agentic AI systems. Dive into the strengths and trade-offs of DeepEval, RAGAS, and TruLens as we unpack how they address multi-step agent evaluation challenges, production readiness, and integration with popular AI toolkits.</p><p><strong>In this episode:</strong></p><p>- Compare DeepEval’s extensive agent-specific metrics and pytest-native integration for development testing</p><p>- Understand RAGAS’s knowledge graph-powered synthetic test generation that slashes test creation time by 90%</p><p>- Discover TruLens’s production-grade observability with hallucination detection via the RAG Triad framework</p><p>- Discuss hybrid evaluation strategies combining these frameworks across the AI lifecycle</p><p>- Learn about real-world deployments in fintech, e-commerce, and enterprise conversational AI</p><p>- Hear expert insights from Keith Bourne on calibration and industry trends</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>DeepEval, RAGAS, TruLens, LangChain, LlamaIndex, LangGraph, OpenTelemetry, Snowflake, Datadog, Cortex AI, DeepTeam</p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction to agentic AI evaluation frameworks</p><p>03:00 - Key metrics and evaluation challenges</p><p>06:30 - Framework architectures and integration</p><p>10:00 - Head-to-head comparison and use cases</p><p>14:00 - Deep technical overview of each framework</p><p>17:30 - Real-world deployments and best practices</p><p>19:30 - Open problems and future directions</p><p><strong>Resources:</strong></p><ol><li data-list="bullet"><span class="ql-ui" contenteditable="false"></span><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e5-agentic-ai-evaluatio" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a> by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e5-agentic-ai-evaluatio" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</li><li data-list="bullet"><span class="ql-ui" contenteditable="false"></span>This podcast is brought to you by <a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e5-agentic-ai-evaluatio" rel="noopener noreferrer" target="_blank">Memriq.ai</a> - AI consultancy and content studio building tools and resources for AI practitioners.</li></ol><br/>]]></description><content:encoded><![CDATA[<p><strong># Evaluating Agentic AI: DeepEval, RAGAS &amp; TruLens Frameworks Compared</strong></p><p>In this episode of Memriq Inference Digest - Engineering Edition, we explore the cutting-edge evaluation frameworks designed for agentic AI systems. Dive into the strengths and trade-offs of DeepEval, RAGAS, and TruLens as we unpack how they address multi-step agent evaluation challenges, production readiness, and integration with popular AI toolkits.</p><p><strong>In this episode:</strong></p><p>- Compare DeepEval’s extensive agent-specific metrics and pytest-native integration for development testing</p><p>- Understand RAGAS’s knowledge graph-powered synthetic test generation that slashes test creation time by 90%</p><p>- Discover TruLens’s production-grade observability with hallucination detection via the RAG Triad framework</p><p>- Discuss hybrid evaluation strategies combining these frameworks across the AI lifecycle</p><p>- Learn about real-world deployments in fintech, e-commerce, and enterprise conversational AI</p><p>- Hear expert insights from Keith Bourne on calibration and industry trends</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>DeepEval, RAGAS, TruLens, LangChain, LlamaIndex, LangGraph, OpenTelemetry, Snowflake, Datadog, Cortex AI, DeepTeam</p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction to agentic AI evaluation frameworks</p><p>03:00 - Key metrics and evaluation challenges</p><p>06:30 - Framework architectures and integration</p><p>10:00 - Head-to-head comparison and use cases</p><p>14:00 - Deep technical overview of each framework</p><p>17:30 - Real-world deployments and best practices</p><p>19:30 - Open problems and future directions</p><p><strong>Resources:</strong></p><ol><li data-list="bullet"><span class="ql-ui" contenteditable="false"></span><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e5-agentic-ai-evaluatio" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a> by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e5-agentic-ai-evaluatio" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</li><li data-list="bullet"><span class="ql-ui" contenteditable="false"></span>This podcast is brought to you by <a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e5-agentic-ai-evaluatio" rel="noopener noreferrer" target="_blank">Memriq.ai</a> - AI consultancy and content studio building tools and resources for AI practitioners.</li></ol><br/>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/agentic-ai-evaluation-frameworks]]></link><guid isPermaLink="false">e7b6706f-dd69-4f45-8ca6-3b7a27297e19</guid><itunes:image href="https://artwork.captivate.fm/8cfa6ba5-e554-4309-b18c-ffff203b500e/artwork-20260104-155255.jpg"/><pubDate>Mon, 05 Jan 2026 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/e7b6706f-dd69-4f45-8ca6-3b7a27297e19.mp3" length="23705804" type="audio/mpeg"/><itunes:duration>19:45</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:season>1</itunes:season><itunes:episode>5</itunes:episode><podcast:episode>5</podcast:episode><podcast:season>1</podcast:season><podcast:transcript url="https://transcripts.captivate.fm/transcript/cd0e5191-273e-4500-a948-ba4fb8df575f/index.html" type="text/html"/></item><item><title>Model Context Protocol: The Universal AI Integration Standard Explained</title><itunes:title>Model Context Protocol: The Universal AI Integration Standard Explained</itunes:title><description><![CDATA[<p>Discover how the Model Context Protocol (MCP) is revolutionizing AI systems integration by simplifying complex multi-tool interactions into a scalable, open standard. In this episode, we unpack MCP’s architecture, adoption by industry leaders, and its impact on engineering workflows.</p><p><strong>In this episode:</strong></p><p>- What MCP is and why it matters for AI/ML engineers and infrastructure teams</p><p>- The M×N integration problem and how MCP reduces it to M+N</p><p>- Core primitives: Tools, Resources, and Prompts, and their roles in MCP</p><p>- Technical deep dive into JSON-RPC 2.0 messaging, transports, and security with OAuth 2.1 + PKCE</p><p>- Comparison of MCP with OpenAI Function Calling, LangChain, and custom REST APIs</p><p>- Real-world adoption, performance metrics, and engineering trade-offs</p><p>- Open challenges including security, authentication, and operational complexity</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>- Model Context Protocol (MCP)</p><p>- JSON-RPC 2.0</p><p>- OAuth 2.1 with PKCE</p><p>- FastMCP Python SDK, MCP TypeScript SDK</p><p>- agentgateway by Solo.io</p><p>- OpenAI Function Calling</p><p>- LangChain</p><p><strong>Timestamps:</strong></p><p>00:00 — Introduction to MCP and episode overview</p><p>02:30 — The M×N integration problem and MCP’s solution</p><p>05:15 — Why MCP adoption is accelerating</p><p>07:00 — MCP architecture and core primitives explained</p><p>10:00 — Head-to-head comparison with alternatives</p><p>12:30 — Under the hood: protocol mechanics and transports</p><p>15:00 — Real-world impact and usage metrics</p><p>17:30 — Challenges and security considerations</p><p>19:00 — Closing thoughts and future outlook</p><p><strong>Resources:</strong></p><ul><li><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e4-model-context-protoc" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a>&nbsp;by&nbsp;<a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e4-model-context-protoc" rel="noopener noreferrer" target="_blank">Keith Bourne</a>&nbsp;- Search for 'Keith Bourne' on Amazon and grab the 2nd edition</li><li>This podcast is brought to you by&nbsp;<a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e4-model-context-protoc" rel="noopener noreferrer" target="_blank">Memriq.ai</a>&nbsp;- AI consultancy and content studio building tools and resources for AI practitioners.</li></ul><br/>]]></description><content:encoded><![CDATA[<p>Discover how the Model Context Protocol (MCP) is revolutionizing AI systems integration by simplifying complex multi-tool interactions into a scalable, open standard. In this episode, we unpack MCP’s architecture, adoption by industry leaders, and its impact on engineering workflows.</p><p><strong>In this episode:</strong></p><p>- What MCP is and why it matters for AI/ML engineers and infrastructure teams</p><p>- The M×N integration problem and how MCP reduces it to M+N</p><p>- Core primitives: Tools, Resources, and Prompts, and their roles in MCP</p><p>- Technical deep dive into JSON-RPC 2.0 messaging, transports, and security with OAuth 2.1 + PKCE</p><p>- Comparison of MCP with OpenAI Function Calling, LangChain, and custom REST APIs</p><p>- Real-world adoption, performance metrics, and engineering trade-offs</p><p>- Open challenges including security, authentication, and operational complexity</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>- Model Context Protocol (MCP)</p><p>- JSON-RPC 2.0</p><p>- OAuth 2.1 with PKCE</p><p>- FastMCP Python SDK, MCP TypeScript SDK</p><p>- agentgateway by Solo.io</p><p>- OpenAI Function Calling</p><p>- LangChain</p><p><strong>Timestamps:</strong></p><p>00:00 — Introduction to MCP and episode overview</p><p>02:30 — The M×N integration problem and MCP’s solution</p><p>05:15 — Why MCP adoption is accelerating</p><p>07:00 — MCP architecture and core primitives explained</p><p>10:00 — Head-to-head comparison with alternatives</p><p>12:30 — Under the hood: protocol mechanics and transports</p><p>15:00 — Real-world impact and usage metrics</p><p>17:30 — Challenges and security considerations</p><p>19:00 — Closing thoughts and future outlook</p><p><strong>Resources:</strong></p><ul><li><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e4-model-context-protoc" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a>&nbsp;by&nbsp;<a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e4-model-context-protoc" rel="noopener noreferrer" target="_blank">Keith Bourne</a>&nbsp;- Search for 'Keith Bourne' on Amazon and grab the 2nd edition</li><li>This podcast is brought to you by&nbsp;<a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e4-model-context-protoc" rel="noopener noreferrer" target="_blank">Memriq.ai</a>&nbsp;- AI consultancy and content studio building tools and resources for AI practitioners.</li></ul><br/>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/model-context-protocol-ai]]></link><guid isPermaLink="false">a5ef7a90-6f8a-4e16-ba83-97d1b7b8c438</guid><itunes:image href="https://artwork.captivate.fm/0a5d6894-5111-4392-8740-9f24ba2d7b21/artwork-20251214-115704.jpg"/><pubDate>Mon, 15 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/a5ef7a90-6f8a-4e16-ba83-97d1b7b8c438.mp3" length="24210284" type="audio/mpeg"/><itunes:duration>20:10</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:season>1</itunes:season><itunes:episode>4</itunes:episode><podcast:episode>4</podcast:episode><podcast:season>1</podcast:season><podcast:transcript url="https://transcripts.captivate.fm/transcript/96ecfe8f-bde7-493d-a3c4-538f6f7c35e4/index.html" type="text/html"/></item><item><title>RAG Evaluation with ragas: Reference-Free Metrics &amp; Monitoring</title><itunes:title>RAG Evaluation with ragas: Reference-Free Metrics &amp; Monitoring</itunes:title><description><![CDATA[<p>Unlock the secrets to evaluating Retrieval-Augmented Generation (RAG) pipelines effectively and efficiently with ragas, the open-source framework that’s transforming AI quality assurance. In this episode, we explore how to implement reference-free evaluation, integrate continuous monitoring into your AI workflows, and optimize for production scale — all through the lens of Keith Bourne’s comprehensive Chapter 9.</p><p><strong>In this episode:</strong></p><p>- Overview of ragas and its reference-free metrics that achieve 95% human agreement on faithfulness scoring</p><p>- Implementation patterns and code walkthroughs for integrating ragas with LangChain, LlamaIndex, and CI/CD pipelines</p><p>- Production monitoring architecture: sampling, async evaluation, aggregation, and alerting</p><p>- Comparison of ragas with other evaluation frameworks like DeepEval and TruLens</p><p>- Strategies for cost optimization and asynchronous evaluation at scale</p><p>- Advanced features: custom domain-specific metrics with AspectCritic and multi-turn evaluation support</p><p><strong>Key tools and technologies mentioned:</strong></p><p>- ragas (Retrieval Augmented Generation Assessment System)</p><p>- LangChain, LlamaIndex</p><p>- LangSmith, LangFuse (observability and evaluation tools)</p><p>- OpenAI GPT-4o, GPT-3.5-turbo, Anthropic Claude, Google Gemini, Ollama</p><p>- Python datasets library</p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction and overview with Keith Bourne</p><p>03:00 - Why reference-free evaluation matters and ragas’s approach</p><p>06:30 - Core metrics: faithfulness, answer relevancy, context precision &amp; recall</p><p>09:00 - Code walkthrough: installation, dataset structure, evaluation calls</p><p>12:00 - Integrations with LangChain, LlamaIndex, and CI/CD workflows</p><p>14:30 - Production monitoring architecture and cost considerations</p><p>17:00 - Advanced metrics and custom domain-specific evaluations</p><p>19:00 - Common pitfalls and testing strategies</p><p>20:30 - Closing thoughts and next steps</p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_podcast3_season1_reference_free_eval" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by Keith Bourne - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Memriq AI: <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">https://Memriq.ai</a></p><p>- ragas website: <a href="https://www.ragas.io/" rel="noopener noreferrer" target="_blank">https://www.ragas.io/</a></p><p>- ragas GitHub repository: <a href="https://github.com/vibrantlabsai/ragas" rel="noopener noreferrer" target="_blank">https://github.com/vibrantlabsai/ragas</a> (for direct access to code and docs)</p><p>Tune in to build more reliable, scalable, and maintainable RAG systems with confidence using open-source evaluation best practices.</p>]]></description><content:encoded><![CDATA[<p>Unlock the secrets to evaluating Retrieval-Augmented Generation (RAG) pipelines effectively and efficiently with ragas, the open-source framework that’s transforming AI quality assurance. In this episode, we explore how to implement reference-free evaluation, integrate continuous monitoring into your AI workflows, and optimize for production scale — all through the lens of Keith Bourne’s comprehensive Chapter 9.</p><p><strong>In this episode:</strong></p><p>- Overview of ragas and its reference-free metrics that achieve 95% human agreement on faithfulness scoring</p><p>- Implementation patterns and code walkthroughs for integrating ragas with LangChain, LlamaIndex, and CI/CD pipelines</p><p>- Production monitoring architecture: sampling, async evaluation, aggregation, and alerting</p><p>- Comparison of ragas with other evaluation frameworks like DeepEval and TruLens</p><p>- Strategies for cost optimization and asynchronous evaluation at scale</p><p>- Advanced features: custom domain-specific metrics with AspectCritic and multi-turn evaluation support</p><p><strong>Key tools and technologies mentioned:</strong></p><p>- ragas (Retrieval Augmented Generation Assessment System)</p><p>- LangChain, LlamaIndex</p><p>- LangSmith, LangFuse (observability and evaluation tools)</p><p>- OpenAI GPT-4o, GPT-3.5-turbo, Anthropic Claude, Google Gemini, Ollama</p><p>- Python datasets library</p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction and overview with Keith Bourne</p><p>03:00 - Why reference-free evaluation matters and ragas’s approach</p><p>06:30 - Core metrics: faithfulness, answer relevancy, context precision &amp; recall</p><p>09:00 - Code walkthrough: installation, dataset structure, evaluation calls</p><p>12:00 - Integrations with LangChain, LlamaIndex, and CI/CD workflows</p><p>14:30 - Production monitoring architecture and cost considerations</p><p>17:00 - Advanced metrics and custom domain-specific evaluations</p><p>19:00 - Common pitfalls and testing strategies</p><p>20:30 - Closing thoughts and next steps</p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_podcast3_season1_reference_free_eval" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by Keith Bourne - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Memriq AI: <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">https://Memriq.ai</a></p><p>- ragas website: <a href="https://www.ragas.io/" rel="noopener noreferrer" target="_blank">https://www.ragas.io/</a></p><p>- ragas GitHub repository: <a href="https://github.com/vibrantlabsai/ragas" rel="noopener noreferrer" target="_blank">https://github.com/vibrantlabsai/ragas</a> (for direct access to code and docs)</p><p>Tune in to build more reliable, scalable, and maintainable RAG systems with confidence using open-source evaluation best practices.</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/rag-evaluation-with-ragas-reference-free-metrics-monitoring]]></link><guid isPermaLink="false">288d30fc-ba3b-4c55-91c4-51b9290b9760</guid><itunes:image href="https://artwork.captivate.fm/51aa787d-1278-48b4-840a-fa5a239ccaaa/artwork-20251214-095742.jpg"/><pubDate>Sun, 14 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/288d30fc-ba3b-4c55-91c4-51b9290b9760.mp3" length="32139404" type="audio/mpeg"/><itunes:duration>26:47</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:season>1</itunes:season><itunes:episode>3</itunes:episode><podcast:episode>3</podcast:episode><podcast:season>1</podcast:season><podcast:transcript url="https://transcripts.captivate.fm/transcript/787a3db6-1024-4553-8339-6f981a0cc8e6/index.html" type="text/html"/></item><item><title>Agent Engineering Unpacked: New Discipline or Just Hype?</title><itunes:title>Agent Engineering Unpacked: New Discipline or Just Hype?</itunes:title><description><![CDATA[<p>Is agent engineering the next big AI discipline or a repackaged buzzword? In this episode, we cut through the hype to explore what agent engineering really means for business leaders navigating AI adoption. From market growth and real-world impact to the critical role of AI memory and the evolving tool landscape, we provide a clear-eyed view to help you make strategic decisions.</p><p><strong>In this episode:</strong></p><p>- The paradox of booming agent engineering markets despite high AI failure rates</p><p>- Why agent engineering is emerging now and what business problems it solves</p><p>- The essential role of AI memory systems and knowledge graphs for real impact</p><p>- Comparing agent engineering frameworks and when to hire agent engineers vs ML engineers</p><p>- Real-world success stories and measurable business payoffs</p><p>- Risks, challenges, and open problems leaders must manage</p><p>Key tools and technologies mentioned: LangChain, LangMem, Mem0, Zep, Memobase, Microsoft AutoGen, Semantic Kernel, CrewAI, OpenAI GPT-4, Anthropic Claude, Google Gemini, Pinecone, Weaviate, Chroma, DeepEval, LangSmith</p><p><strong>Timestamps:</strong></p><p>00:00 – Introduction &amp; Why Agent Engineering Matters</p><p>03:45 – Market Overview &amp; The Paradox of AI Agent Performance</p><p>07:30 – Why Now: Technology and Talent Trends Driving Adoption</p><p>11:15 – The Big Picture: Managing AI Unpredictability</p><p>14:00 – The Memory Imperative: Transforming AI Agents</p><p>17:00 – Knowledge Graphs &amp; Domain Expertise</p><p>19:30 – Framework Landscape &amp; When to Hire Agent Engineers</p><p>22:45 – How Agent Engineering Works: A Simplified View</p><p>26:00 – Real-World Payoffs &amp; Business Impact</p><p>29:15 – Reality Check: Risks and Limitations</p><p>32:30 – Agent Engineering In the Wild: Industry Use Cases</p><p>35:00 – Tech Battle: Agent Engineers vs ML Engineers</p><p>38:00 – Toolbox for Leaders: Strategic Considerations</p><p>41:00 – Book Spotlight &amp; Sponsor Message</p><p>43:00 – Open Problems &amp; Future Outlook</p><p>45:00 – Final Words &amp; Closing Remarks</p><p><strong>Resources:</strong></p><ul><li><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=leadership-s1-e2-agent-engineering-hy" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a>&nbsp;by&nbsp;<a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=leadership-s1-e2-agent-engineering-hy" rel="noopener noreferrer" target="_blank">Keith Bourne</a>&nbsp;- Search for 'Keith Bourne' on Amazon and grab the 2nd edition</li><li>This podcast is brought to you by&nbsp;<a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=leadership-s1-e2-agent-engineering-hy" rel="noopener noreferrer" target="_blank">Memriq.ai</a>&nbsp;- AI consultancy and content studio building tools and resources for AI practitioners.</li></ul><br/><p>Thanks for tuning into Memriq Inference Digest - Engineering Edition. Stay curious and keep building!</p>]]></description><content:encoded><![CDATA[<p>Is agent engineering the next big AI discipline or a repackaged buzzword? In this episode, we cut through the hype to explore what agent engineering really means for business leaders navigating AI adoption. From market growth and real-world impact to the critical role of AI memory and the evolving tool landscape, we provide a clear-eyed view to help you make strategic decisions.</p><p><strong>In this episode:</strong></p><p>- The paradox of booming agent engineering markets despite high AI failure rates</p><p>- Why agent engineering is emerging now and what business problems it solves</p><p>- The essential role of AI memory systems and knowledge graphs for real impact</p><p>- Comparing agent engineering frameworks and when to hire agent engineers vs ML engineers</p><p>- Real-world success stories and measurable business payoffs</p><p>- Risks, challenges, and open problems leaders must manage</p><p>Key tools and technologies mentioned: LangChain, LangMem, Mem0, Zep, Memobase, Microsoft AutoGen, Semantic Kernel, CrewAI, OpenAI GPT-4, Anthropic Claude, Google Gemini, Pinecone, Weaviate, Chroma, DeepEval, LangSmith</p><p><strong>Timestamps:</strong></p><p>00:00 – Introduction &amp; Why Agent Engineering Matters</p><p>03:45 – Market Overview &amp; The Paradox of AI Agent Performance</p><p>07:30 – Why Now: Technology and Talent Trends Driving Adoption</p><p>11:15 – The Big Picture: Managing AI Unpredictability</p><p>14:00 – The Memory Imperative: Transforming AI Agents</p><p>17:00 – Knowledge Graphs &amp; Domain Expertise</p><p>19:30 – Framework Landscape &amp; When to Hire Agent Engineers</p><p>22:45 – How Agent Engineering Works: A Simplified View</p><p>26:00 – Real-World Payoffs &amp; Business Impact</p><p>29:15 – Reality Check: Risks and Limitations</p><p>32:30 – Agent Engineering In the Wild: Industry Use Cases</p><p>35:00 – Tech Battle: Agent Engineers vs ML Engineers</p><p>38:00 – Toolbox for Leaders: Strategic Considerations</p><p>41:00 – Book Spotlight &amp; Sponsor Message</p><p>43:00 – Open Problems &amp; Future Outlook</p><p>45:00 – Final Words &amp; Closing Remarks</p><p><strong>Resources:</strong></p><ul><li><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=leadership-s1-e2-agent-engineering-hy" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a>&nbsp;by&nbsp;<a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=leadership-s1-e2-agent-engineering-hy" rel="noopener noreferrer" target="_blank">Keith Bourne</a>&nbsp;- Search for 'Keith Bourne' on Amazon and grab the 2nd edition</li><li>This podcast is brought to you by&nbsp;<a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=leadership-s1-e2-agent-engineering-hy" rel="noopener noreferrer" target="_blank">Memriq.ai</a>&nbsp;- AI consultancy and content studio building tools and resources for AI practitioners.</li></ul><br/><p>Thanks for tuning into Memriq Inference Digest - Engineering Edition. Stay curious and keep building!</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/agent-engineering-discussion]]></link><guid isPermaLink="false">571b5d9f-5ad8-4342-8142-e5fe668f8ea5</guid><itunes:image href="https://artwork.captivate.fm/099aa518-5ec2-48db-9b46-16e0998acd9a/artwork-20251213-143042.jpg"/><pubDate>Sat, 13 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/571b5d9f-5ad8-4342-8142-e5fe668f8ea5.mp3" length="36050444" type="audio/mpeg"/><itunes:duration>30:02</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:season>1</itunes:season><itunes:episode>2</itunes:episode><podcast:episode>2</podcast:episode><podcast:season>1</podcast:season><podcast:transcript url="https://transcripts.captivate.fm/transcript/1406835b-6b79-44d9-8026-6d714f03e938/index.html" type="text/html"/></item><item><title>Why Your AI Is Failing: The NLU Paradigm Shift CTOs Can’t Ignore</title><itunes:title>Why Your AI Is Failing: The NLU Paradigm Shift CTOs Can’t Ignore</itunes:title><description><![CDATA[<p>Are your AI initiatives stalling in production? This episode uncovers the critical architectural shift brought by the Natural Language Understanding (NLU) layer and why treating AI as just another feature is setting CTOs up for failure. Learn how rethinking your entire stack—from closed-world deterministic workflows to open-world AI-driven orchestration—is essential to unlock real business value.</p><p><strong>In this episode:</strong></p><p>- Understand the fundamental difference between traditional deterministic web apps and AI-powered conversational interfaces</p><p>- Explore the pivotal role of the NLU layer as the "brain" that dynamically interprets, prioritizes, and routes user intents</p><p>- Discover why adding an orchestrator component bridges the gap between probabilistic AI reasoning and deterministic backend execution</p><p>- Dive into multi-intent handling, partial understanding, and strategies for graceful fallback and out-of-scope requests</p><p>- Compare architectural approaches and learn best practices for building production-grade AI chatbots</p><p>- Hear about real-world deployments and open challenges facing AI/ML engineers and infrastructure teams</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>- Large Language Models (LLMs)</p><p>- Structured function calling APIs</p><p>- Conversational AI orchestrators</p><p>- 99-intents fallback pattern</p><p>- Semantic caching and episodic memory</p><p><strong>Timestamps:</strong></p><p>00:00 – Introduction &amp; Why This Matters</p><p>03:30 – The NLU Paradigm Shift Explained</p><p>07:45 – The Orchestrator: Bridging AI and Backend</p><p>11:20 – Handling Multi-Intent &amp; Partial Understanding</p><p>14:10 – Turning Fallbacks into Opportunities</p><p>16:50 – Architectural Comparisons &amp; Best Practices</p><p>19:30 – Real-World Deployments &amp; Open Problems</p><p>22:15 – Final Takeaways &amp; Closing</p><p><strong>Resources:</strong></p><ul><li><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e1-nlu-paradigm-shift-a" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a>&nbsp;by&nbsp;<a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e1-nlu-paradigm-shift-a" rel="noopener noreferrer" target="_blank">Keith Bourne</a>&nbsp;- Search for 'Keith Bourne' on Amazon and grab the 2nd edition</li><li>This podcast is brought to you by&nbsp;<a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e1-nlu-paradigm-shift-a" rel="noopener noreferrer" target="_blank">Memriq.ai</a>&nbsp;- AI consultancy and content studio building tools and resources for AI practitioners.</li></ul><br/>]]></description><content:encoded><![CDATA[<p>Are your AI initiatives stalling in production? This episode uncovers the critical architectural shift brought by the Natural Language Understanding (NLU) layer and why treating AI as just another feature is setting CTOs up for failure. Learn how rethinking your entire stack—from closed-world deterministic workflows to open-world AI-driven orchestration—is essential to unlock real business value.</p><p><strong>In this episode:</strong></p><p>- Understand the fundamental difference between traditional deterministic web apps and AI-powered conversational interfaces</p><p>- Explore the pivotal role of the NLU layer as the "brain" that dynamically interprets, prioritizes, and routes user intents</p><p>- Discover why adding an orchestrator component bridges the gap between probabilistic AI reasoning and deterministic backend execution</p><p>- Dive into multi-intent handling, partial understanding, and strategies for graceful fallback and out-of-scope requests</p><p>- Compare architectural approaches and learn best practices for building production-grade AI chatbots</p><p>- Hear about real-world deployments and open challenges facing AI/ML engineers and infrastructure teams</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>- Large Language Models (LLMs)</p><p>- Structured function calling APIs</p><p>- Conversational AI orchestrators</p><p>- 99-intents fallback pattern</p><p>- Semantic caching and episodic memory</p><p><strong>Timestamps:</strong></p><p>00:00 – Introduction &amp; Why This Matters</p><p>03:30 – The NLU Paradigm Shift Explained</p><p>07:45 – The Orchestrator: Bridging AI and Backend</p><p>11:20 – Handling Multi-Intent &amp; Partial Understanding</p><p>14:10 – Turning Fallbacks into Opportunities</p><p>16:50 – Architectural Comparisons &amp; Best Practices</p><p>19:30 – Real-World Deployments &amp; Open Problems</p><p>22:15 – Final Takeaways &amp; Closing</p><p><strong>Resources:</strong></p><ul><li><a href="https://a.co/d/4h3kgub?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=rag-book-2e&amp;utm_content=engineering-s1-e1-nlu-paradigm-shift-a" rel="noopener noreferrer" target="_blank">"Unlocking Data with Generative AI and RAG"</a>&nbsp;by&nbsp;<a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=keith-linkedin&amp;utm_content=engineering-s1-e1-nlu-paradigm-shift-a" rel="noopener noreferrer" target="_blank">Keith Bourne</a>&nbsp;- Search for 'Keith Bourne' on Amazon and grab the 2nd edition</li><li>This podcast is brought to you by&nbsp;<a href="https://memriq.ai/?utm_source=memriq-podcast&amp;utm_medium=show-notes&amp;utm_campaign=memriq-website&amp;utm_content=engineering-s1-e1-nlu-paradigm-shift-a" rel="noopener noreferrer" target="_blank">Memriq.ai</a>&nbsp;- AI consultancy and content studio building tools and resources for AI practitioners.</li></ul><br/>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/nlu-paradigm-shift-ai-failures]]></link><guid isPermaLink="false">03056b9c-4356-4156-9f94-ce9d838fd594</guid><itunes:image href="https://artwork.captivate.fm/7898eb04-67f1-42a4-bb88-8d02b9285a0f/artwork-20251213-135410.jpg"/><pubDate>Sat, 13 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/03056b9c-4356-4156-9f94-ce9d838fd594.mp3" length="46240364" type="audio/mpeg"/><itunes:duration>38:32</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:season>1</itunes:season><itunes:episode>1</itunes:episode><podcast:episode>1</podcast:episode><podcast:season>1</podcast:season><podcast:transcript url="https://transcripts.captivate.fm/transcript/79339239-1f9d-499e-9ceb-5556577d713c/index.html" type="text/html"/></item><item><title>Advanced RAG with Complete Memory Integration (Chapter 19)</title><itunes:title>Advanced RAG with Complete Memory Integration (Chapter 19)</itunes:title><description><![CDATA[<p>Unlock the next level of Retrieval-Augmented Generation with full memory integration in AI agents.  In the previous 3 episodes, we secretly built up what amounts to a 4-part series on agentic memory.  This is the final piece of that 4-part series that pulls it ALL together. </p><p>In this episode, we explore how combining episodic, semantic, and procedural memories via the CoALA architecture and LangMem library transforms static retrieval systems into continuously learning, adaptive AI.</p><p>This also concludes our book series, highlighting ALL of the chapters of the 2nd edition of "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>.  If you want to dive even deeper into these topics and even try out extensive code labs, search for 'Keith Bourne' on Amazon and grab the 2nd edition today!</p><p>In this episode:</p><p>- How CoALAAgent unifies multiple memory types for dynamic AI behavior</p><p>- Trade-offs between LangMem’s prompt_memory, gradient, and metaprompt algorithms</p><p>- Architectural patterns for modular and scalable AI agent development</p><p>- Real-world metrics demonstrating continuous procedural strategy learning</p><p>- Challenges around data quality, metric design, and domain agent engineering</p><p>- Practical advice for building safe, adaptive AI agents in production</p><p>Key tools &amp; technologies: CoALAAgent, LangMem library, GPT models, hierarchical memory scopes</p><p>Timestamps:</p><p>0:00 Intro &amp; guest welcome</p><p>3:30 Why integrating episodic, semantic &amp; procedural memory matters</p><p>7:15 The CoALA architecture and hierarchical learning scopes</p><p>10:00 Comparing procedural learning algorithms in LangMem</p><p>13:30 Behind the scenes: memory integration pipeline</p><p>16:00 Real-world impact &amp; procedural strategy success metrics</p><p>18:30 Challenges in deploying memory-integrated RAG systems</p><p>20:00 Practical engineering tips &amp; closing thoughts</p><p><br></p><p>Resources:</p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Memriq AI: <a href="https://memriq.ai" rel="noopener noreferrer" target="_blank">https://memriq.ai</a></p>]]></description><content:encoded><![CDATA[<p>Unlock the next level of Retrieval-Augmented Generation with full memory integration in AI agents.  In the previous 3 episodes, we secretly built up what amounts to a 4-part series on agentic memory.  This is the final piece of that 4-part series that pulls it ALL together. </p><p>In this episode, we explore how combining episodic, semantic, and procedural memories via the CoALA architecture and LangMem library transforms static retrieval systems into continuously learning, adaptive AI.</p><p>This also concludes our book series, highlighting ALL of the chapters of the 2nd edition of "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>.  If you want to dive even deeper into these topics and even try out extensive code labs, search for 'Keith Bourne' on Amazon and grab the 2nd edition today!</p><p>In this episode:</p><p>- How CoALAAgent unifies multiple memory types for dynamic AI behavior</p><p>- Trade-offs between LangMem’s prompt_memory, gradient, and metaprompt algorithms</p><p>- Architectural patterns for modular and scalable AI agent development</p><p>- Real-world metrics demonstrating continuous procedural strategy learning</p><p>- Challenges around data quality, metric design, and domain agent engineering</p><p>- Practical advice for building safe, adaptive AI agents in production</p><p>Key tools &amp; technologies: CoALAAgent, LangMem library, GPT models, hierarchical memory scopes</p><p>Timestamps:</p><p>0:00 Intro &amp; guest welcome</p><p>3:30 Why integrating episodic, semantic &amp; procedural memory matters</p><p>7:15 The CoALA architecture and hierarchical learning scopes</p><p>10:00 Comparing procedural learning algorithms in LangMem</p><p>13:30 Behind the scenes: memory integration pipeline</p><p>16:00 Real-world impact &amp; procedural strategy success metrics</p><p>18:30 Challenges in deploying memory-integrated RAG systems</p><p>20:00 Practical engineering tips &amp; closing thoughts</p><p><br></p><p>Resources:</p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Memriq AI: <a href="https://memriq.ai" rel="noopener noreferrer" target="_blank">https://memriq.ai</a></p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/advanced-rag-with-complete-memory-integration-chapter-19]]></link><guid isPermaLink="false">e0e48749-63c4-4450-90fb-ce2b6605741d</guid><itunes:image href="https://artwork.captivate.fm/63a1f60f-5fb2-4862-b161-176f2915e85b/artwork-20251212-123916.jpg"/><pubDate>Fri, 12 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/e0e48749-63c4-4450-90fb-ce2b6605741d.mp3" length="20707244" type="audio/mpeg"/><itunes:duration>17:15</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>17</itunes:episode><podcast:episode>17</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/a591a8e7-f591-4028-92ca-83da88400b3a/index.html" type="text/html"/></item><item><title>Procedural Memory for RAG:  Deep Dive with LangMem (Chapter 18)</title><itunes:title>Procedural Memory for RAG:  Deep Dive with LangMem (Chapter 18)</itunes:title><description><![CDATA[<p>Unlock the power of procedural memory to transform your Retrieval-Augmented Generation (RAG) agents into autonomous learners. In this episode, we explore how LangMem leverages hierarchical learning scopes to enable AI agents that continuously adapt and improve from their interactions — cutting down manual tuning and boosting real-world performance.</p><p><strong>In this episode:</strong></p><p>- Why procedural memory is a game changer for RAG systems and the challenges it addresses</p><p>- How LangMem integrates with LangChain and OpenAI GPT-4.1-mini to implement procedural memory</p><p>- The architecture patterns behind hierarchical namespaces and momentum-based feedback loops</p><p>- Trade-offs between traditional RAG and LangMem’s procedural memory approach</p><p>- Real-world applications across finance, healthcare, education, and customer service</p><p>- Practical engineering tips, monitoring best practices, and open problems in procedural memory</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>- LangMem</p><p>- LangChain</p><p>- Pydantic</p><p>- OpenAI GPT-4.1-mini</p><p><br></p><p><strong>Timestamps:</strong></p><p>0:00 - Introduction &amp; overview</p><p>2:30 - Why procedural memory matters now</p><p>5:15 - Core concepts &amp; hierarchical learning scopes</p><p>8:45 - LangMem architecture &amp; domain interface</p><p>12:00 - Trade-offs: Traditional RAG vs LangMem</p><p>14:30 - Real-world use cases &amp; impact</p><p>17:00 - Engineering best practices &amp; pitfalls</p><p>19:30 - Open challenges &amp; future outlook</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Memriq AI: <a href="https://memriq.ai" rel="noopener noreferrer" target="_blank">https://memriq.ai</a></p>]]></description><content:encoded><![CDATA[<p>Unlock the power of procedural memory to transform your Retrieval-Augmented Generation (RAG) agents into autonomous learners. In this episode, we explore how LangMem leverages hierarchical learning scopes to enable AI agents that continuously adapt and improve from their interactions — cutting down manual tuning and boosting real-world performance.</p><p><strong>In this episode:</strong></p><p>- Why procedural memory is a game changer for RAG systems and the challenges it addresses</p><p>- How LangMem integrates with LangChain and OpenAI GPT-4.1-mini to implement procedural memory</p><p>- The architecture patterns behind hierarchical namespaces and momentum-based feedback loops</p><p>- Trade-offs between traditional RAG and LangMem’s procedural memory approach</p><p>- Real-world applications across finance, healthcare, education, and customer service</p><p>- Practical engineering tips, monitoring best practices, and open problems in procedural memory</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>- LangMem</p><p>- LangChain</p><p>- Pydantic</p><p>- OpenAI GPT-4.1-mini</p><p><br></p><p><strong>Timestamps:</strong></p><p>0:00 - Introduction &amp; overview</p><p>2:30 - Why procedural memory matters now</p><p>5:15 - Core concepts &amp; hierarchical learning scopes</p><p>8:45 - LangMem architecture &amp; domain interface</p><p>12:00 - Trade-offs: Traditional RAG vs LangMem</p><p>14:30 - Real-world use cases &amp; impact</p><p>17:00 - Engineering best practices &amp; pitfalls</p><p>19:30 - Open challenges &amp; future outlook</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Memriq AI: <a href="https://memriq.ai" rel="noopener noreferrer" target="_blank">https://memriq.ai</a></p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/procedural-memory-for-rag-deep-dive-with-langmem-chapter-18]]></link><guid isPermaLink="false">e5dccc70-16e6-4bd4-b2e2-f9d8cd748793</guid><itunes:image href="https://artwork.captivate.fm/2a8cb35c-d477-42c3-beda-ec18c5b7bf39/artwork-20251212-120331.jpg"/><pubDate>Fri, 12 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/e5dccc70-16e6-4bd4-b2e2-f9d8cd748793.mp3" length="21770924" type="audio/mpeg"/><itunes:duration>18:09</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>16</itunes:episode><podcast:episode>16</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/5ef7b4a2-dcba-4cb2-919f-64bb4768431e/index.html" type="text/html"/></item><item><title>RAG-Based Agentic Memory: Code Perspective (Chapter 17)</title><itunes:title>RAG-Based Agentic Memory: Code Perspective (Chapter 17)</itunes:title><description><![CDATA[<p>Unlock how Retrieval-Augmented Generation (RAG) enables AI agents to remember, learn, and personalize over time. In this episode, we explore Chapter 17 of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>," focusing on implementing agentic memory with the CoALA framework. From episodic and semantic memory distinctions to real-world engineering trade-offs, this discussion is packed with practical insights for AI/ML engineers and infrastructure experts.</p><p><strong>In this episode:</strong></p><p>- Understand the difference between episodic and semantic memory and their roles in AI agents</p><p>- Explore how vector databases like ChromaDB power fast, scalable memory retrieval</p><p>- Dive into the architecture and code walkthrough using CoALA, LangChain, LangGraph, and OpenAI APIs</p><p>- Discuss engineering challenges including validation, latency, and system complexity</p><p>- Hear from author <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> on the foundational importance of agentic memory</p><p>- Review real-world applications and open problems shaping the future of memory-augmented AI</p><p><strong>Key tools and technologies mentioned:</strong></p><p>- CoALA framework</p><p>- LangChain &amp; LangGraph</p><p>- ChromaDB vector database</p><p>- OpenAI API (embeddings and LLMs)</p><p>- python-dotenv</p><p>- Pydantic models</p><p><strong>Timestamps:</strong></p><p>0:00 - Introduction &amp; Episode Overview</p><p>2:30 - The Concept of Agentic Memory: Episodic vs Semantic</p><p>6:00 - Vector Databases and Retrieval-Augmented Generation (RAG)</p><p>9:30 - Coding Agentic Memory: Frameworks and Workflow</p><p>13:00 - Engineering Trade-offs and Validation Challenges</p><p>16:00 - Real-World Applications and Use Cases</p><p>18:30 - Open Problems and Future Directions</p><p>20:00 - Closing Thoughts and Resources</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit Memriq AI at <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">https://Memriq.ai</a> for more AI engineering deep dives and resources</p>]]></description><content:encoded><![CDATA[<p>Unlock how Retrieval-Augmented Generation (RAG) enables AI agents to remember, learn, and personalize over time. In this episode, we explore Chapter 17 of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>," focusing on implementing agentic memory with the CoALA framework. From episodic and semantic memory distinctions to real-world engineering trade-offs, this discussion is packed with practical insights for AI/ML engineers and infrastructure experts.</p><p><strong>In this episode:</strong></p><p>- Understand the difference between episodic and semantic memory and their roles in AI agents</p><p>- Explore how vector databases like ChromaDB power fast, scalable memory retrieval</p><p>- Dive into the architecture and code walkthrough using CoALA, LangChain, LangGraph, and OpenAI APIs</p><p>- Discuss engineering challenges including validation, latency, and system complexity</p><p>- Hear from author <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> on the foundational importance of agentic memory</p><p>- Review real-world applications and open problems shaping the future of memory-augmented AI</p><p><strong>Key tools and technologies mentioned:</strong></p><p>- CoALA framework</p><p>- LangChain &amp; LangGraph</p><p>- ChromaDB vector database</p><p>- OpenAI API (embeddings and LLMs)</p><p>- python-dotenv</p><p>- Pydantic models</p><p><strong>Timestamps:</strong></p><p>0:00 - Introduction &amp; Episode Overview</p><p>2:30 - The Concept of Agentic Memory: Episodic vs Semantic</p><p>6:00 - Vector Databases and Retrieval-Augmented Generation (RAG)</p><p>9:30 - Coding Agentic Memory: Frameworks and Workflow</p><p>13:00 - Engineering Trade-offs and Validation Challenges</p><p>16:00 - Real-World Applications and Use Cases</p><p>18:30 - Open Problems and Future Directions</p><p>20:00 - Closing Thoughts and Resources</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit Memriq AI at <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">https://Memriq.ai</a> for more AI engineering deep dives and resources</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/rag-based-agentic-memory-understanding-the-code-from-a-leader-perspective-chapter-17]]></link><guid isPermaLink="false">28b19204-589c-4898-81b6-943c6011e90b</guid><itunes:image href="https://artwork.captivate.fm/f783b5b4-76af-48f7-82b3-410aaeffa86e/artwork-20251212-103510.jpg"/><pubDate>Fri, 12 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/28b19204-589c-4898-81b6-943c6011e90b.mp3" length="20140844" type="audio/mpeg"/><itunes:duration>16:47</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>15</itunes:episode><podcast:episode>15</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/5c8809dc-60bf-4cc1-b802-7dc89538aeae/index.html" type="text/html"/></item><item><title>Agentic Memory:  Stateful RAG and AI Agents (Chapter 16)</title><itunes:title>Agentic Memory:  Stateful RAG and AI Agents (Chapter 16)</itunes:title><description><![CDATA[<p>Unlock the future of AI agents with agentic memory — a transformative approach that extends Retrieval-Augmented Generation (RAG) by incorporating persistent, evolving memories. In this episode, we explore how stateful intelligence turns stateless LLMs into adaptive, personalized agents capable of learning over time.</p><p><strong>In this episode:</strong></p><p>- Understand the CoALA framework dividing memory into episodic, semantic, procedural, and working types</p><p>- Explore key tools like Mem0, LangMem, Zep, Graphiti, LangChain, and Neo4j for implementing agentic memory</p><p>- Dive into practical architectural patterns, memory curation strategies, and trade-offs for real-world AI systems</p><p>- Hear from <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>, author of <em>*</em><a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank"><em>Unlocking Data with Generative AI and RAG</em></a><em>*</em>, sharing insider insights and code lab highlights</p><p>- Discuss latency, accuracy improvements, and engineering challenges in scaling stateful AI agents</p><p>- Review real-world applications across finance, healthcare, education, and customer support</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>Mem0, LangMem, Zep, Graphiti, LangChain, Neo4j, Pinecone, Weaviate, Airflow, Temporal</p><p><br></p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction &amp; Episode Overview</p><p>02:15 - What is Agentic Memory and Why It Matters</p><p>06:10 - The CoALA Cognitive Architecture Explained</p><p>09:30 - Comparing Memory Implementations: Mem0, LangMem, Graphiti</p><p>13:00 - Deep Dive: Memory Curation and Background Pipelines</p><p>16:00 - Performance Metrics &amp; Real-World Impact</p><p>18:30 - Challenges &amp; Open Problems in Agentic Memory</p><p>20:00 - Closing Thoughts &amp; Resources</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for more AI engineering deep dives and resources</p>]]></description><content:encoded><![CDATA[<p>Unlock the future of AI agents with agentic memory — a transformative approach that extends Retrieval-Augmented Generation (RAG) by incorporating persistent, evolving memories. In this episode, we explore how stateful intelligence turns stateless LLMs into adaptive, personalized agents capable of learning over time.</p><p><strong>In this episode:</strong></p><p>- Understand the CoALA framework dividing memory into episodic, semantic, procedural, and working types</p><p>- Explore key tools like Mem0, LangMem, Zep, Graphiti, LangChain, and Neo4j for implementing agentic memory</p><p>- Dive into practical architectural patterns, memory curation strategies, and trade-offs for real-world AI systems</p><p>- Hear from <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>, author of <em>*</em><a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank"><em>Unlocking Data with Generative AI and RAG</em></a><em>*</em>, sharing insider insights and code lab highlights</p><p>- Discuss latency, accuracy improvements, and engineering challenges in scaling stateful AI agents</p><p>- Review real-world applications across finance, healthcare, education, and customer support</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>Mem0, LangMem, Zep, Graphiti, LangChain, Neo4j, Pinecone, Weaviate, Airflow, Temporal</p><p><br></p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction &amp; Episode Overview</p><p>02:15 - What is Agentic Memory and Why It Matters</p><p>06:10 - The CoALA Cognitive Architecture Explained</p><p>09:30 - Comparing Memory Implementations: Mem0, LangMem, Graphiti</p><p>13:00 - Deep Dive: Memory Curation and Background Pipelines</p><p>16:00 - Performance Metrics &amp; Real-World Impact</p><p>18:30 - Challenges &amp; Open Problems in Agentic Memory</p><p>20:00 - Closing Thoughts &amp; Resources</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for more AI engineering deep dives and resources</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/agentic-memory-stateful-rag-and-ai-agents-chapter-16]]></link><guid isPermaLink="false">d0081443-ae8b-47d1-8faa-5816cfb78211</guid><itunes:image href="https://artwork.captivate.fm/22118dc4-ca15-4aa0-a956-44a03b46b3b2/artwork-20251212-100404.jpg"/><pubDate>Fri, 12 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/d0081443-ae8b-47d1-8faa-5816cfb78211.mp3" length="23340524" type="audio/mpeg"/><itunes:duration>19:27</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>14</itunes:episode><podcast:episode>14</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/59cc2ea7-281b-4ffc-b3b7-4586accbb5ce/index.html" type="text/html"/></item><item><title>Semantic Caches: Scaling AI with Smarter Caching (Chapter 15)</title><itunes:title>Semantic Caches: Scaling AI with Smarter Caching (Chapter 15)</itunes:title><description><![CDATA[<p>emantic caches are transforming how AI systems handle costly reasoning by intelligently reusing prior agent workflows to slash latency and inference costs. In this episode, we unpack Chapter 15 of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>," exploring the architectures, trade-offs, and practical engineering of semantic caches for production AI.</p><p><strong>In this episode:</strong></p><p>- What semantic caches are and why they reduce AI inference latency by up to 100x</p><p>- Core techniques: vector embeddings, entity masking, and CrossEncoder verification</p><p>- Comparing semantic cache variants and fallback strategies for robust performance</p><p>- Under-the-hood implementation details using ChromaDB, sentence-transformers, and CrossEncoder</p><p>- Real-world use cases across finance, customer support, and enterprise AI assistants</p><p>- Key challenges: tuning thresholds, cache eviction, and maintaining precision in production</p><p><strong>Key tools and technologies mentioned:</strong></p><p>- ChromaDB vector database</p><p>- Sentence-transformers embedding models (e.g., all-mpnet-base-v2)</p><p>- CrossEncoder models for verification</p><p>- Regex-based entity masking</p><p>- Adaptive similarity thresholding</p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction and episode overview</p><p>02:30 - What are semantic caches and why now?</p><p>06:15 - Core architecture: embedding, masking, and verification</p><p>10:00 - Semantic cache variants and fallback approaches</p><p>13:30 - Implementation walkthrough using Python and ChromaDB</p><p>16:00 - Real-world applications and performance metrics</p><p>18:30 - Open problems and engineering challenges</p><p>19:30 - Final thoughts and book spotlight</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Memriq AI: <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">https://Memriq.ai</a></p>]]></description><content:encoded><![CDATA[<p>emantic caches are transforming how AI systems handle costly reasoning by intelligently reusing prior agent workflows to slash latency and inference costs. In this episode, we unpack Chapter 15 of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>," exploring the architectures, trade-offs, and practical engineering of semantic caches for production AI.</p><p><strong>In this episode:</strong></p><p>- What semantic caches are and why they reduce AI inference latency by up to 100x</p><p>- Core techniques: vector embeddings, entity masking, and CrossEncoder verification</p><p>- Comparing semantic cache variants and fallback strategies for robust performance</p><p>- Under-the-hood implementation details using ChromaDB, sentence-transformers, and CrossEncoder</p><p>- Real-world use cases across finance, customer support, and enterprise AI assistants</p><p>- Key challenges: tuning thresholds, cache eviction, and maintaining precision in production</p><p><strong>Key tools and technologies mentioned:</strong></p><p>- ChromaDB vector database</p><p>- Sentence-transformers embedding models (e.g., all-mpnet-base-v2)</p><p>- CrossEncoder models for verification</p><p>- Regex-based entity masking</p><p>- Adaptive similarity thresholding</p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction and episode overview</p><p>02:30 - What are semantic caches and why now?</p><p>06:15 - Core architecture: embedding, masking, and verification</p><p>10:00 - Semantic cache variants and fallback approaches</p><p>13:30 - Implementation walkthrough using Python and ChromaDB</p><p>16:00 - Real-world applications and performance metrics</p><p>18:30 - Open problems and engineering challenges</p><p>19:30 - Final thoughts and book spotlight</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Memriq AI: <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">https://Memriq.ai</a></p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/semantic-caches-scaling-ai-with-smarter-caching-chapter-15]]></link><guid isPermaLink="false">441f91b3-e026-469b-ac9b-24d07fe8d6ec</guid><itunes:image href="https://artwork.captivate.fm/cd55f984-129b-4938-b4eb-776e8026d762/artwork-20251212-091700.jpg"/><pubDate>Fri, 12 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/441f91b3-e026-469b-ac9b-24d07fe8d6ec.mp3" length="24082604" type="audio/mpeg"/><itunes:duration>20:04</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>13</itunes:episode><podcast:episode>13</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/25e0fb3c-b1f0-4ad4-a596-2a8c0921cf5e/index.html" type="text/html"/></item><item><title>Graph-Based RAG: Hybrid Embeddings &amp; Explainable AI (Chapter 14)</title><itunes:title>Graph-Based RAG: Hybrid Embeddings &amp; Explainable AI (Chapter 14)</itunes:title><description><![CDATA[<p>Unlock the power of graph-based Retrieval-Augmented Generation (RAG) in this technical deep dive featuring insights from Chapter 14 of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>'s "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>." Discover how combining knowledge graphs with LLMs using hybrid embeddings and explicit graph traversal can dramatically improve multi-hop reasoning accuracy and explainability.</p><p>In this episode:</p><p>- Explore ontology design and graph ingestion workflows using Protégé, RDF, and Neo4j</p><p>- Understand the advantages of hybrid embeddings over vector-only approaches</p><p>- Learn why Python static dictionaries significantly boost LLM multi-hop reasoning accuracy</p><p>- Discuss architecture trade-offs between ontology-based and cyclical graph RAG systems</p><p>- Review real-world production considerations, scalability challenges, and tooling best practices</p><p>- Hear directly from author <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> about building explainable and reliable AI pipelines</p><p>Key tools and technologies mentioned:</p><p>- Protégé for ontology creation</p><p>- RDF triples and rdflib for data parsing</p><p>- Neo4j graph database with Cypher queries</p><p>- Sentence-Transformers (all-MiniLM-L6-v2) for embedding generation</p><p>- FAISS for vector similarity search</p><p>- LangChain for orchestration</p><p>- OpenAI chat models</p><p>- python-dotenv for secrets management</p><p><br></p><p>Timestamps:</p><p>00:00 - Introduction &amp; episode overview</p><p>02:30 - Surprising results: Python dicts vs natural language for KG representation</p><p>05:45 - Why graph-based RAG matters now: tech readiness &amp; industry demand</p><p>08:15 - Architecture walkthrough: from ontology to LLM prompt input</p><p>12:00 - Comparing ontology-based vs cyclical graph RAG approaches</p><p>15:00 - Under the hood: building the pipeline step-by-step</p><p>18:30 - Real-world results, scaling challenges, and practical tips</p><p>21:00 - Closing thoughts and next steps</p><p><br></p><p>Resources:</p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit Memriq AI at <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">https://Memriq.ai</a> for more AI engineering insights and tools</p>]]></description><content:encoded><![CDATA[<p>Unlock the power of graph-based Retrieval-Augmented Generation (RAG) in this technical deep dive featuring insights from Chapter 14 of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>'s "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>." Discover how combining knowledge graphs with LLMs using hybrid embeddings and explicit graph traversal can dramatically improve multi-hop reasoning accuracy and explainability.</p><p>In this episode:</p><p>- Explore ontology design and graph ingestion workflows using Protégé, RDF, and Neo4j</p><p>- Understand the advantages of hybrid embeddings over vector-only approaches</p><p>- Learn why Python static dictionaries significantly boost LLM multi-hop reasoning accuracy</p><p>- Discuss architecture trade-offs between ontology-based and cyclical graph RAG systems</p><p>- Review real-world production considerations, scalability challenges, and tooling best practices</p><p>- Hear directly from author <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> about building explainable and reliable AI pipelines</p><p>Key tools and technologies mentioned:</p><p>- Protégé for ontology creation</p><p>- RDF triples and rdflib for data parsing</p><p>- Neo4j graph database with Cypher queries</p><p>- Sentence-Transformers (all-MiniLM-L6-v2) for embedding generation</p><p>- FAISS for vector similarity search</p><p>- LangChain for orchestration</p><p>- OpenAI chat models</p><p>- python-dotenv for secrets management</p><p><br></p><p>Timestamps:</p><p>00:00 - Introduction &amp; episode overview</p><p>02:30 - Surprising results: Python dicts vs natural language for KG representation</p><p>05:45 - Why graph-based RAG matters now: tech readiness &amp; industry demand</p><p>08:15 - Architecture walkthrough: from ontology to LLM prompt input</p><p>12:00 - Comparing ontology-based vs cyclical graph RAG approaches</p><p>15:00 - Under the hood: building the pipeline step-by-step</p><p>18:30 - Real-world results, scaling challenges, and practical tips</p><p>21:00 - Closing thoughts and next steps</p><p><br></p><p>Resources:</p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit Memriq AI at <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">https://Memriq.ai</a> for more AI engineering insights and tools</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/graph-based-rag-hybrid-embeddings-explainable-ai-chapter-14]]></link><guid isPermaLink="false">a3da22eb-e5a1-479a-b255-9cdd6c8b209f</guid><itunes:image href="https://artwork.captivate.fm/c8e00cf5-4dd6-4431-9045-caedac6934ad/artwork-20251212-084354.jpg"/><pubDate>Fri, 12 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/a3da22eb-e5a1-479a-b255-9cdd6c8b209f.mp3" length="26071724" type="audio/mpeg"/><itunes:duration>21:44</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>12</itunes:episode><podcast:episode>12</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/c5e25976-df88-4cd2-bf31-19602189b297/index.html" type="text/html"/></item><item><title>Ontology-Based Knowledge Engineering for Graphs (Chapter 13)</title><itunes:title>Ontology-Based Knowledge Engineering for Graphs (Chapter 13)</itunes:title><description><![CDATA[<p>Ontologies are the semantic backbone that enable AI systems to reason precisely over complex domain knowledge, far beyond what vector embeddings alone can achieve. In this episode, we explore ontology-based knowledge engineering for graph-backed AI, featuring insights from <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>'s Chapter 13 of <em>*</em><a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank"><em>Unlocking Data with Generative AI and RAG</em></a><em>*</em>. Learn how ontologies empower multi-hop reasoning, improve explainability, and support scalable, production-grade AI systems.</p><p><strong>In this episode:</strong></p><p>- The fundamentals of ontologies, OWL, RDFS, and Protégé for building semantically rich knowledge graphs</p><p>- How ontology-based reasoning enhances retrieval-augmented generation (RAG) pipelines with precise domain constraints</p><p>- Practical tooling and workflows: from ontology authoring and validation to Neo4j graph integration</p><p>- Trade-offs between expressivity, performance, and maintainability in ontology engineering</p><p>- Real-world use cases across finance, healthcare, and compliance where ontologies enable trustworthy AI</p><p>- Open challenges and future directions in ontology automation, scalability, and hybrid AI systems</p><p><strong>Key tools and technologies mentioned:</strong></p><p>- Protégé (ontology authoring and reasoning)</p><p>- OWL 2 DL (Web Ontology Language for expressive domain modeling)</p><p>- RDFS and SKOS (vocabularies for annotation and lightweight semantics)</p><p>- Neo4j (graph database for knowledge graph storage and traversal)</p><p>- OWL reasoners (Pellet, HermiT, Fact++)</p><p><br></p><p><strong>Timestamps:</strong></p><p>00:00 – Introduction and episode overview</p><p>02:30 – Why ontologies matter now in AI and RAG</p><p>05:15 – Ontology basics: classes, properties, and logical constraints</p><p>08:00 – Tooling walkthrough: Protégé, OWL, Neo4j integration</p><p>11:45 – Performance and production considerations</p><p>14:30 – Real-world applications and case studies</p><p>17:00 – Technical trade-offs and best practices</p><p>19:15 – Open problems and future outlook</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> – Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit Memriq AI for tools and resources: <a href="https://memriq.ai" rel="noopener noreferrer" target="_blank">https://memriq.ai</a></p>]]></description><content:encoded><![CDATA[<p>Ontologies are the semantic backbone that enable AI systems to reason precisely over complex domain knowledge, far beyond what vector embeddings alone can achieve. In this episode, we explore ontology-based knowledge engineering for graph-backed AI, featuring insights from <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>'s Chapter 13 of <em>*</em><a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank"><em>Unlocking Data with Generative AI and RAG</em></a><em>*</em>. Learn how ontologies empower multi-hop reasoning, improve explainability, and support scalable, production-grade AI systems.</p><p><strong>In this episode:</strong></p><p>- The fundamentals of ontologies, OWL, RDFS, and Protégé for building semantically rich knowledge graphs</p><p>- How ontology-based reasoning enhances retrieval-augmented generation (RAG) pipelines with precise domain constraints</p><p>- Practical tooling and workflows: from ontology authoring and validation to Neo4j graph integration</p><p>- Trade-offs between expressivity, performance, and maintainability in ontology engineering</p><p>- Real-world use cases across finance, healthcare, and compliance where ontologies enable trustworthy AI</p><p>- Open challenges and future directions in ontology automation, scalability, and hybrid AI systems</p><p><strong>Key tools and technologies mentioned:</strong></p><p>- Protégé (ontology authoring and reasoning)</p><p>- OWL 2 DL (Web Ontology Language for expressive domain modeling)</p><p>- RDFS and SKOS (vocabularies for annotation and lightweight semantics)</p><p>- Neo4j (graph database for knowledge graph storage and traversal)</p><p>- OWL reasoners (Pellet, HermiT, Fact++)</p><p><br></p><p><strong>Timestamps:</strong></p><p>00:00 – Introduction and episode overview</p><p>02:30 – Why ontologies matter now in AI and RAG</p><p>05:15 – Ontology basics: classes, properties, and logical constraints</p><p>08:00 – Tooling walkthrough: Protégé, OWL, Neo4j integration</p><p>11:45 – Performance and production considerations</p><p>14:30 – Real-world applications and case studies</p><p>17:00 – Technical trade-offs and best practices</p><p>19:15 – Open problems and future outlook</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> – Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit Memriq AI for tools and resources: <a href="https://memriq.ai" rel="noopener noreferrer" target="_blank">https://memriq.ai</a></p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/ontology-based-knowledge-engineering-for-graphs-chapter-13]]></link><guid isPermaLink="false">694c9bb4-4828-4e87-b8d0-2c12c0c273e5</guid><itunes:image href="https://artwork.captivate.fm/06d90729-bf96-4de4-9038-bf35b5d31c2b/artwork-20251212-081315.jpg"/><pubDate>Fri, 12 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/694c9bb4-4828-4e87-b8d0-2c12c0c273e5.mp3" length="22939244" type="audio/mpeg"/><itunes:duration>19:07</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>11</itunes:episode><podcast:episode>11</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/a360c900-a9ef-4664-ad73-5adf646c5d8d/index.html" type="text/html"/></item><item><title>Agentic RAG &amp; LangGraph: Next-Gen AI Orchestration (Chapter 12)</title><itunes:title>Agentic RAG &amp; LangGraph: Next-Gen AI Orchestration (Chapter 12)</itunes:title><description><![CDATA[<p>Unlock the next evolution of Retrieval-Augmented Generation in this episode of Memriq Inference Digest – Engineering Edition. We explore how combining AI agents with LangGraph's graph-based orchestration transforms brittle linear RAG pipelines into dynamic, multi-step reasoning systems that self-correct and scale.</p><p><strong>In this episode:</strong></p><p>- Understand the shift from linear RAG to agentic workflows with dynamic tool invocation and query refinement loops</p><p>- Dive into LangGraph’s graph orchestration model for managing complex, conditional control flows with state persistence</p><p>- Explore the synergy between LangChain tools, ChatOpenAI, and third-party APIs like TavilySearch for multi-source retrieval</p><p>- Get under the hood with code patterns including AgentState design, conditional edges, and streaming LLM calls</p><p>- Hear from <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>, author of “<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>,” on practical lessons and architectural best practices</p><p>- Discuss trade-offs in latency, complexity, debugging, and production readiness for agentic RAG systems</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>- LangGraph (StateGraph, ToolNode)</p><p>- LangChain (retriever tools, bind_tools)</p><p>- ChatOpenAI (streaming LLM interface)</p><p>- Pydantic (structured output validation)</p><p>- TavilySearch (live web search API)</p><p><br></p><p><strong>Timestamps:</strong></p><p>0:00 – Intro and episode overview</p><p>2:15 – Why agentic RAG and LangGraph matter now</p><p>5:30 – Big picture: graph-based agent orchestration</p><p>8:45 – Head-to-head: linear RAG vs. agentic RAG</p><p>11:20 – Under the hood: building agent workflows with LangGraph</p><p>14:50 – Payoff: performance gains and multi-source retrieval</p><p>17:10 – Reality check: challenges &amp; pitfalls in agent design</p><p>19:00 – Real-world applications and case studies</p><p>21:30 – Toolbox tips for engineers</p><p>23:45 – Book spotlight &amp; final thoughts</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> – Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://memriq.ai" rel="noopener noreferrer" target="_blank">https://memriq.ai</a> for more AI deep dives, practical guides, and research breakdowns</p><p><br></p><p>Thanks for listening to Memriq Inference Digest. Stay tuned for more engineering insights into the evolving AI landscape.</p>]]></description><content:encoded><![CDATA[<p>Unlock the next evolution of Retrieval-Augmented Generation in this episode of Memriq Inference Digest – Engineering Edition. We explore how combining AI agents with LangGraph's graph-based orchestration transforms brittle linear RAG pipelines into dynamic, multi-step reasoning systems that self-correct and scale.</p><p><strong>In this episode:</strong></p><p>- Understand the shift from linear RAG to agentic workflows with dynamic tool invocation and query refinement loops</p><p>- Dive into LangGraph’s graph orchestration model for managing complex, conditional control flows with state persistence</p><p>- Explore the synergy between LangChain tools, ChatOpenAI, and third-party APIs like TavilySearch for multi-source retrieval</p><p>- Get under the hood with code patterns including AgentState design, conditional edges, and streaming LLM calls</p><p>- Hear from <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>, author of “<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>,” on practical lessons and architectural best practices</p><p>- Discuss trade-offs in latency, complexity, debugging, and production readiness for agentic RAG systems</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>- LangGraph (StateGraph, ToolNode)</p><p>- LangChain (retriever tools, bind_tools)</p><p>- ChatOpenAI (streaming LLM interface)</p><p>- Pydantic (structured output validation)</p><p>- TavilySearch (live web search API)</p><p><br></p><p><strong>Timestamps:</strong></p><p>0:00 – Intro and episode overview</p><p>2:15 – Why agentic RAG and LangGraph matter now</p><p>5:30 – Big picture: graph-based agent orchestration</p><p>8:45 – Head-to-head: linear RAG vs. agentic RAG</p><p>11:20 – Under the hood: building agent workflows with LangGraph</p><p>14:50 – Payoff: performance gains and multi-source retrieval</p><p>17:10 – Reality check: challenges &amp; pitfalls in agent design</p><p>19:00 – Real-world applications and case studies</p><p>21:30 – Toolbox tips for engineers</p><p>23:45 – Book spotlight &amp; final thoughts</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> – Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://memriq.ai" rel="noopener noreferrer" target="_blank">https://memriq.ai</a> for more AI deep dives, practical guides, and research breakdowns</p><p><br></p><p>Thanks for listening to Memriq Inference Digest. Stay tuned for more engineering insights into the evolving AI landscape.</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/agentic-rag-langgraph-next-gen-ai-orchestration-chapter-12]]></link><guid isPermaLink="false">0cb275aa-cf97-4993-8aff-f299ac96235d</guid><itunes:image href="https://artwork.captivate.fm/3709c991-d35a-463f-9092-9a80b0e2e3d2/artwork-20251212-002330.jpg"/><pubDate>Fri, 12 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/0cb275aa-cf97-4993-8aff-f299ac96235d.mp3" length="25506764" type="audio/mpeg"/><itunes:duration>21:15</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>10</itunes:episode><podcast:episode>10</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/ff87edc0-1c77-419c-affe-1ad4cc2df57a/index.html" type="text/html"/></item><item><title>Using LangChain to Get More from RAG (Chapter 11)</title><itunes:title>Using LangChain to Get More from RAG (Chapter 11)</itunes:title><description><![CDATA[<p>Unlock the full potential of Retrieval-Augmented Generation (RAG) with LangChain’s modular components in this episode of Memriq Inference Digest — Engineering Edition. We dive deep into <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Chapter 11</a> of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">book</a>, exploring how document loaders, semantic text splitters, and structured output parsers can transform your RAG pipelines for better data ingestion, retrieval relevance, and reliable downstream automation.</p><p><strong>In this episode:</strong></p><p>- Explore LangChain’s diverse document loaders for PDFs, HTML, Word docs, and JSON</p><p>- Understand semantic chunking with RecursiveCharacterTextSplitter versus naive splitting</p><p>- Learn about structured output parsing using JsonOutputParser and Pydantic models</p><p>- Compare tooling trade-offs for building scalable and maintainable RAG systems</p><p>- Hear real-world use cases across enterprise knowledge bases, customer support, and compliance</p><p>- Get practical engineering tips to optimize pipeline latency, metadata hygiene, and robustness</p><p><strong>Key tools &amp; technologies:</strong></p><p>- LangChain document loaders (PyPDF2, BSHTMLLoader, Docx2txtLoader, JSONLoader)</p><p>- RecursiveCharacterTextSplitter</p><p>- Output parsers: StrOutputParser, JsonOutputParser with Pydantic</p><p>- OpenAI text-embedding-ada-002</p><p><br></p><p><strong>Timestamps:</strong></p><p>00:00 – Introduction and guest welcome</p><p>02:30 – The power of LangChain’s modular components</p><p>06:00 – Why LangChain’s approach matters now</p><p>08:30 – Core RAG pipeline architecture breakdown</p><p>11:30 – Tool comparisons: loaders, splitters, parsers</p><p>14:30 – Under the hood walkthrough</p><p>17:00 – Real-world applications and engineering trade-offs</p><p>19:30 – Closing thoughts and resources</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit Memriq.ai for more AI engineering deep dives and resources</p>]]></description><content:encoded><![CDATA[<p>Unlock the full potential of Retrieval-Augmented Generation (RAG) with LangChain’s modular components in this episode of Memriq Inference Digest — Engineering Edition. We dive deep into <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Chapter 11</a> of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">book</a>, exploring how document loaders, semantic text splitters, and structured output parsers can transform your RAG pipelines for better data ingestion, retrieval relevance, and reliable downstream automation.</p><p><strong>In this episode:</strong></p><p>- Explore LangChain’s diverse document loaders for PDFs, HTML, Word docs, and JSON</p><p>- Understand semantic chunking with RecursiveCharacterTextSplitter versus naive splitting</p><p>- Learn about structured output parsing using JsonOutputParser and Pydantic models</p><p>- Compare tooling trade-offs for building scalable and maintainable RAG systems</p><p>- Hear real-world use cases across enterprise knowledge bases, customer support, and compliance</p><p>- Get practical engineering tips to optimize pipeline latency, metadata hygiene, and robustness</p><p><strong>Key tools &amp; technologies:</strong></p><p>- LangChain document loaders (PyPDF2, BSHTMLLoader, Docx2txtLoader, JSONLoader)</p><p>- RecursiveCharacterTextSplitter</p><p>- Output parsers: StrOutputParser, JsonOutputParser with Pydantic</p><p>- OpenAI text-embedding-ada-002</p><p><br></p><p><strong>Timestamps:</strong></p><p>00:00 – Introduction and guest welcome</p><p>02:30 – The power of LangChain’s modular components</p><p>06:00 – Why LangChain’s approach matters now</p><p>08:30 – Core RAG pipeline architecture breakdown</p><p>11:30 – Tool comparisons: loaders, splitters, parsers</p><p>14:30 – Under the hood walkthrough</p><p>17:00 – Real-world applications and engineering trade-offs</p><p>19:30 – Closing thoughts and resources</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit Memriq.ai for more AI engineering deep dives and resources</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/using-langchain-to-get-more-from-rag-chapter-11]]></link><guid isPermaLink="false">22fe056e-edee-43bf-9b29-ba7086b1d3c5</guid><itunes:image href="https://artwork.captivate.fm/c5e581b5-b07a-4833-90d9-aa2b3341050f/artwork-20251211-235705.jpg"/><pubDate>Fri, 12 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/22fe056e-edee-43bf-9b29-ba7086b1d3c5.mp3" length="28155404" type="audio/mpeg"/><itunes:duration>23:28</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>9</itunes:episode><podcast:episode>9</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/d814712f-07e7-4b3b-ade4-6f5e402aae54/index.html" type="text/html"/></item><item><title>Key RAG Components in LangChain (Chapter 10)</title><itunes:title>Key RAG Components in LangChain (Chapter 10)</itunes:title><description><![CDATA[<p>Unlock the inner workings of Retrieval-Augmented Generation (RAG) pipelines using LangChain in this episode of Memriq Inference Digest - Engineering Edition. We bring insights directly from <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>, author of '<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>,' as we explore modular vector stores, retrievers, and LLM integrations critical for building scalable, flexible AI systems.</p><p><strong>In this episode:</strong></p><p>- Explore LangChain’s modular architecture for building RAG pipelines</p><p>- Compare popular vector stores: Chroma, FAISS, Weaviate, and Pinecone</p><p>- Understand retriever strategies: BM25, dense, and ensemble approaches</p><p>- Dive into LLM integrations like OpenAI’s ChatOpenAI and Together AI’s ChatTogether</p><p>- Discuss engineering trade-offs, GPU acceleration, and production considerations</p><p>- Highlight real-world use cases and challenges in scaling retrieval</p><p>Key tools and technologies mentioned:</p><p>- LangChain framework</p><p>- Vector stores: Chroma, FAISS, Weaviate, Pinecone</p><p>- Retrievers: BM25, Dense, Ensemble Retriever</p><p>- LLMs: OpenAI ChatOpenAI, Together AI ChatTogether</p><p>- FAISS GPU acceleration</p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction &amp; episode overview</p><p>02:15 - LangChain modularity and design philosophy</p><p>05:30 - Vector store comparisons and scale trade-offs</p><p>09:00 - Retriever types and ensemble approaches</p><p>12:30 - Under the hood: pipeline walkthrough</p><p>15:00 - Performance metrics and latency improvements</p><p>17:00 - Real-world applications and challenges</p><p>19:00 - Final thoughts and book spotlight</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for AI infrastructure deep dives, practical guides, and research breakdowns</p><p><br></p><p>Thanks for tuning in to Memriq Inference Digest - Engineering Edition. Stay curious and keep building!</p>]]></description><content:encoded><![CDATA[<p>Unlock the inner workings of Retrieval-Augmented Generation (RAG) pipelines using LangChain in this episode of Memriq Inference Digest - Engineering Edition. We bring insights directly from <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>, author of '<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>,' as we explore modular vector stores, retrievers, and LLM integrations critical for building scalable, flexible AI systems.</p><p><strong>In this episode:</strong></p><p>- Explore LangChain’s modular architecture for building RAG pipelines</p><p>- Compare popular vector stores: Chroma, FAISS, Weaviate, and Pinecone</p><p>- Understand retriever strategies: BM25, dense, and ensemble approaches</p><p>- Dive into LLM integrations like OpenAI’s ChatOpenAI and Together AI’s ChatTogether</p><p>- Discuss engineering trade-offs, GPU acceleration, and production considerations</p><p>- Highlight real-world use cases and challenges in scaling retrieval</p><p>Key tools and technologies mentioned:</p><p>- LangChain framework</p><p>- Vector stores: Chroma, FAISS, Weaviate, Pinecone</p><p>- Retrievers: BM25, Dense, Ensemble Retriever</p><p>- LLMs: OpenAI ChatOpenAI, Together AI ChatTogether</p><p>- FAISS GPU acceleration</p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction &amp; episode overview</p><p>02:15 - LangChain modularity and design philosophy</p><p>05:30 - Vector store comparisons and scale trade-offs</p><p>09:00 - Retriever types and ensemble approaches</p><p>12:30 - Under the hood: pipeline walkthrough</p><p>15:00 - Performance metrics and latency improvements</p><p>17:00 - Real-world applications and challenges</p><p>19:00 - Final thoughts and book spotlight</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for AI infrastructure deep dives, practical guides, and research breakdowns</p><p><br></p><p>Thanks for tuning in to Memriq Inference Digest - Engineering Edition. Stay curious and keep building!</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/key-rag-components-in-langchain-chapter-10]]></link><guid isPermaLink="false">9a56bc05-e74e-4ed3-a763-55ad947d3c32</guid><itunes:image href="https://artwork.captivate.fm/e647e03f-013e-4ca6-b249-51ed352ed253/artwork-20251211-233256.jpg"/><pubDate>Thu, 11 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/9a56bc05-e74e-4ed3-a763-55ad947d3c32.mp3" length="21970124" type="audio/mpeg"/><itunes:duration>18:18</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>8</itunes:episode><podcast:episode>8</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/83e61414-820e-4815-bee7-b2ed0835d1c9/index.html" type="text/html"/></item><item><title>Evaluating RAG: Quantitative Metrics and Visual Insights (Chapter 9)</title><itunes:title>Evaluating RAG: Quantitative Metrics and Visual Insights (Chapter 9)</itunes:title><description><![CDATA[<p>In this episode of Memriq Inference Digest — Engineering Edition, we dive deep into rigorous evaluation strategies for Retrieval-Augmented Generation (RAG) systems. Drawing from <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Chapter 9</a> of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">book</a>, we explore how quantitative metrics and visualizations help AI engineers optimize retrieval and generation performance while managing cost and complexity.</p><p><strong>In this episode:</strong></p><p>- Why continuous, multi-metric evaluation is critical for RAG pipelines post-deployment</p><p>- Comparing dense vector similarity search versus hybrid search with real metric trade-offs</p><p>- Automating synthetic ground truth generation using LLMs wrapped in LangChain</p><p>- Building modular, scalable evaluation pipelines with ragas and visualization tools</p><p>- Practical challenges like cost management, dataset size limitations, and the role of human evaluation</p><p>- Real-world use cases in finance, research, and customer support that benefit from rigorous evaluation</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>- ragas (open-source RAG evaluation framework)</p><p>- LangChain (model and embedding wrappers)</p><p>- matplotlib and pandas (data visualization and manipulation)</p><p>- ChatOpenAI (LLM for generation and evaluation)</p><p><br></p><p><strong>Timestamps:</strong></p><p>0:00 – Introduction and episode overview</p><p>2:30 – The importance of continuous RAG evaluation</p><p>5:15 – Hybrid vs dense similarity search: metric comparisons</p><p>9:00 – Under the hood: ragas evaluation pipeline and LangChain wrappers</p><p>13:00 – Visualizing RAG metrics for actionable insights</p><p>16:00 – Practical limitations and balancing cost with thoroughness</p><p>18:30 – Real-world RAG evaluation examples</p><p>21:00 – Open challenges and future directions</p><p>23:30 – Final thoughts and book spotlight</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for more AI engineering deep-dives, tools, and resources</p>]]></description><content:encoded><![CDATA[<p>In this episode of Memriq Inference Digest — Engineering Edition, we dive deep into rigorous evaluation strategies for Retrieval-Augmented Generation (RAG) systems. Drawing from <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Chapter 9</a> of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">book</a>, we explore how quantitative metrics and visualizations help AI engineers optimize retrieval and generation performance while managing cost and complexity.</p><p><strong>In this episode:</strong></p><p>- Why continuous, multi-metric evaluation is critical for RAG pipelines post-deployment</p><p>- Comparing dense vector similarity search versus hybrid search with real metric trade-offs</p><p>- Automating synthetic ground truth generation using LLMs wrapped in LangChain</p><p>- Building modular, scalable evaluation pipelines with ragas and visualization tools</p><p>- Practical challenges like cost management, dataset size limitations, and the role of human evaluation</p><p>- Real-world use cases in finance, research, and customer support that benefit from rigorous evaluation</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>- ragas (open-source RAG evaluation framework)</p><p>- LangChain (model and embedding wrappers)</p><p>- matplotlib and pandas (data visualization and manipulation)</p><p>- ChatOpenAI (LLM for generation and evaluation)</p><p><br></p><p><strong>Timestamps:</strong></p><p>0:00 – Introduction and episode overview</p><p>2:30 – The importance of continuous RAG evaluation</p><p>5:15 – Hybrid vs dense similarity search: metric comparisons</p><p>9:00 – Under the hood: ragas evaluation pipeline and LangChain wrappers</p><p>13:00 – Visualizing RAG metrics for actionable insights</p><p>16:00 – Practical limitations and balancing cost with thoroughness</p><p>18:30 – Real-world RAG evaluation examples</p><p>21:00 – Open challenges and future directions</p><p>23:30 – Final thoughts and book spotlight</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for more AI engineering deep-dives, tools, and resources</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/evaluating-rag-quantitative-metrics-and-visual-insights-chapter-9]]></link><guid isPermaLink="false">a803850f-29cb-4e79-90db-68cabba657ce</guid><itunes:image href="https://artwork.captivate.fm/955873fc-ae05-4fb4-bfde-d3137ca23de9/artwork-20251211-230215.jpg"/><pubDate>Thu, 11 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/a803850f-29cb-4e79-90db-68cabba657ce.mp3" length="24268844" type="audio/mpeg"/><itunes:duration>20:13</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>7</itunes:episode><podcast:episode>7</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/468c6ac7-d77a-402e-b79b-9629583c6dfb/index.html" type="text/html"/></item><item><title>Similarity Searching with Vectors (Chapter 8)</title><itunes:title>Similarity Searching with Vectors (Chapter 8)</itunes:title><description><![CDATA[<p>Unlock the power of similarity search with vectors in this episode of Memriq Inference Digest – Engineering Edition. We explore how dense and sparse vector techniques combine to enable scalable, accurate semantic retrieval for AI systems, inspired by <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Chapter 8</a> of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">book</a>. Join us and special guest <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> as we unpack the engineering trade-offs, indexing algorithms, hybrid search strategies, and real-world applications that make vector search foundational in modern AI workflows.</p><p><strong>In this episode:</strong></p><p>- The fundamentals of representing data as high-dimensional embeddings and retrieving nearest neighbors</p><p>- How hybrid search fuses dense semantic embeddings with sparse keyword vectors to boost relevance</p><p>- Deep dive into Approximate Nearest Neighbor algorithms like HNSW for billion-scale indexing</p><p>- Practical considerations between open-source models and managed vector stores</p><p>- Engineering tips on tuning ANN parameters, persistence, and combining retrieval results with Reciprocal Rank Fusion</p><p>- Real-world use cases in enterprise search, recommendation engines, and retrieval-augmented generation systems</p><p><strong>Key tools and technologies mentioned:</strong></p><p>- sentence_transformers (e.g., all-mpnet-base-v2)</p><p>- BM25Retriever</p><p>- LangChain and Chroma</p><p>- FAISS, HNSW, ANNOY</p><p>- Reciprocal Rank Fusion (RRF)</p><p>- Pinecone, Weaviate, Google Vertex AI Vector Search</p><p><br></p><p><strong>Timestamps:</strong></p><p>0:00 - Introduction and episode overview</p><p>2:00 - The power of hybrid search: dense + sparse vectors</p><p>5:30 - ANN algorithms and indexing techniques (HNSW, LSH)</p><p>9:00 - Trade-offs: open-source embeddings vs commercial APIs</p><p>11:30 - Reciprocal Rank Fusion and ranking strategies</p><p>14:00 - Engineering challenges: persistence, tuning, and latency</p><p>16:30 - Real-world applications and production system considerations</p><p>19:00 - Final thoughts and resources</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for advanced AI engineering guides and resources</p><p><br></p><p>Thanks for tuning into Memriq Inference Digest – Engineering Edition. Stay sharp, and see you next time!</p>]]></description><content:encoded><![CDATA[<p>Unlock the power of similarity search with vectors in this episode of Memriq Inference Digest – Engineering Edition. We explore how dense and sparse vector techniques combine to enable scalable, accurate semantic retrieval for AI systems, inspired by <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Chapter 8</a> of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">book</a>. Join us and special guest <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> as we unpack the engineering trade-offs, indexing algorithms, hybrid search strategies, and real-world applications that make vector search foundational in modern AI workflows.</p><p><strong>In this episode:</strong></p><p>- The fundamentals of representing data as high-dimensional embeddings and retrieving nearest neighbors</p><p>- How hybrid search fuses dense semantic embeddings with sparse keyword vectors to boost relevance</p><p>- Deep dive into Approximate Nearest Neighbor algorithms like HNSW for billion-scale indexing</p><p>- Practical considerations between open-source models and managed vector stores</p><p>- Engineering tips on tuning ANN parameters, persistence, and combining retrieval results with Reciprocal Rank Fusion</p><p>- Real-world use cases in enterprise search, recommendation engines, and retrieval-augmented generation systems</p><p><strong>Key tools and technologies mentioned:</strong></p><p>- sentence_transformers (e.g., all-mpnet-base-v2)</p><p>- BM25Retriever</p><p>- LangChain and Chroma</p><p>- FAISS, HNSW, ANNOY</p><p>- Reciprocal Rank Fusion (RRF)</p><p>- Pinecone, Weaviate, Google Vertex AI Vector Search</p><p><br></p><p><strong>Timestamps:</strong></p><p>0:00 - Introduction and episode overview</p><p>2:00 - The power of hybrid search: dense + sparse vectors</p><p>5:30 - ANN algorithms and indexing techniques (HNSW, LSH)</p><p>9:00 - Trade-offs: open-source embeddings vs commercial APIs</p><p>11:30 - Reciprocal Rank Fusion and ranking strategies</p><p>14:00 - Engineering challenges: persistence, tuning, and latency</p><p>16:30 - Real-world applications and production system considerations</p><p>19:00 - Final thoughts and resources</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for advanced AI engineering guides and resources</p><p><br></p><p>Thanks for tuning into Memriq Inference Digest – Engineering Edition. Stay sharp, and see you next time!</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/similarity-searching-with-vectors-chapter-8]]></link><guid isPermaLink="false">1be06bca-03f4-42d3-96ce-e8120f1b493b</guid><itunes:image href="https://artwork.captivate.fm/32998bea-6d9c-4239-84fb-e7a3e659f993/artwork-20251211-223557.jpg"/><pubDate>Thu, 11 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/1be06bca-03f4-42d3-96ce-e8120f1b493b.mp3" length="24685004" type="audio/mpeg"/><itunes:duration>20:34</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>6</itunes:episode><podcast:episode>6</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/e7d49165-7344-4a02-bf8e-d6da9b17d677/index.html" type="text/html"/></item><item><title>Vectors &amp; Vector Stores in RAG (Chapter 7)</title><itunes:title>Vectors &amp; Vector Stores in RAG (Chapter 7)</itunes:title><description><![CDATA[<p>Unlock the core infrastructure powering retrieval-augmented generation (RAG) systems in this technical deep dive. We explore how vector embeddings and vector stores work together to enable fast, scalable, and semantically rich retrieval for LLMs, drawing insights directly from <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Chapter 7</a> of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">book</a>.</p><p><strong>In this episode:</strong></p><p>- Understand the role of high-dimensional vectors and vector stores in powering RAG</p><p>- Compare embedding models like OpenAIEmbeddings, BERT, and Doc2Vec</p><p>- Explore vector store technologies including Chroma, Milvus, Pinecone, and pgvector</p><p>- Deep dive into indexing algorithms like HNSW and adaptive retrieval techniques such as Matryoshka embeddings</p><p>- Discuss architectural trade-offs for production-ready RAG systems</p><p>- Hear real-world applications and operational challenges from embedding compatibility to scaling</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>OpenAIEmbeddings, BERT, Doc2Vec, Chroma, Milvus, Pinecone, pgvector, LangChain, HNSW, Matryoshka embeddings</p><p><br></p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction to vectors and vector stores in RAG</p><p>02:15 - Why vectors are the backbone of retrieval-augmented generation</p><p>05:40 - Embedding models: trade-offs and use cases</p><p>09:10 - Vector stores and indexing: Chroma, Milvus, Pinecone, pgvector</p><p>13:00 - Under the hood: indexing algorithms and adaptive retrieval</p><p>16:20 - Real-world deployments and architectural trade-offs</p><p>18:40 - Open challenges and best practices</p><p>20:30 - Final thoughts and book recommendation</p><p><br></p><p>Resources:</p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for more AI practitioner tools, resources, and deep dives</p>]]></description><content:encoded><![CDATA[<p>Unlock the core infrastructure powering retrieval-augmented generation (RAG) systems in this technical deep dive. We explore how vector embeddings and vector stores work together to enable fast, scalable, and semantically rich retrieval for LLMs, drawing insights directly from <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Chapter 7</a> of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">book</a>.</p><p><strong>In this episode:</strong></p><p>- Understand the role of high-dimensional vectors and vector stores in powering RAG</p><p>- Compare embedding models like OpenAIEmbeddings, BERT, and Doc2Vec</p><p>- Explore vector store technologies including Chroma, Milvus, Pinecone, and pgvector</p><p>- Deep dive into indexing algorithms like HNSW and adaptive retrieval techniques such as Matryoshka embeddings</p><p>- Discuss architectural trade-offs for production-ready RAG systems</p><p>- Hear real-world applications and operational challenges from embedding compatibility to scaling</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>OpenAIEmbeddings, BERT, Doc2Vec, Chroma, Milvus, Pinecone, pgvector, LangChain, HNSW, Matryoshka embeddings</p><p><br></p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction to vectors and vector stores in RAG</p><p>02:15 - Why vectors are the backbone of retrieval-augmented generation</p><p>05:40 - Embedding models: trade-offs and use cases</p><p>09:10 - Vector stores and indexing: Chroma, Milvus, Pinecone, pgvector</p><p>13:00 - Under the hood: indexing algorithms and adaptive retrieval</p><p>16:20 - Real-world deployments and architectural trade-offs</p><p>18:40 - Open challenges and best practices</p><p>20:30 - Final thoughts and book recommendation</p><p><br></p><p>Resources:</p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for more AI practitioner tools, resources, and deep dives</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/vectors-vector-stores-in-rag-chapter-7]]></link><guid isPermaLink="false">22460fda-ed3f-402c-a524-07461a676806</guid><itunes:image href="https://artwork.captivate.fm/e99ea878-e7ec-4e00-b177-bc6da57c2062/artwork-20251211-221038.jpg"/><pubDate>Thu, 11 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/22460fda-ed3f-402c-a524-07461a676806.mp3" length="23671724" type="audio/mpeg"/><itunes:duration>19:44</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>5</itunes:episode><podcast:episode>5</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/eddc1fa6-c34f-400f-9df2-2551e130ede9/index.html" type="text/html"/></item><item><title>Interfacing RAG with Gradio: Rapid Prototyping (Chapter 6)</title><itunes:title>Interfacing RAG with Gradio: Rapid Prototyping (Chapter 6)</itunes:title><description><![CDATA[<p>Unlock the power of retrieval-augmented generation (RAG) by integrating it seamlessly with Gradio. In this episode, we explore how Gradio simplifies building interactive RAG applications, enabling AI engineers to prototype and share demos quickly without complex frontend coding.</p><p><strong>In this episode:</strong></p><p>- Discover how Gradio’s `demo.launch(share=True)` command spins up shareable RAG UIs in minutes</p><p>- Understand environment setup challenges like nested asyncio event loops and uvloop conflicts</p><p>- Compare Gradio’s rapid prototyping advantages with production-ready custom frontends</p><p>- Learn deployment options including Hugging Face Spaces and LangChain integration</p><p>- Hear insider insights from <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>, author of “<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>”</p><p>- Discuss real-world use cases, security trade-offs, and scaling considerations</p><p><strong>Key tools &amp; technologies: </strong>Gradio, RAG pipelines, LangChain, Hugging Face Spaces, Python asyncio, nest_asyncio, uvloop</p><p><br></p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction and episode overview</p><p>02:15 - What is Gradio and why it matters for RAG</p><p>05:30 - Rapid prototyping with `demo.launch(share=True)`</p><p>08:45 - Environment quirks: asyncio loops and uvloop</p><p>11:20 - Architectural trade-offs: Gradio vs custom frontends</p><p>14:10 - Deployment strategies and hosting on Hugging Face Spaces</p><p>17:00 - Security considerations and production readiness</p><p>19:15 - Closing thoughts and resources</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for more AI engineering deep dives and practical guides</p>]]></description><content:encoded><![CDATA[<p>Unlock the power of retrieval-augmented generation (RAG) by integrating it seamlessly with Gradio. In this episode, we explore how Gradio simplifies building interactive RAG applications, enabling AI engineers to prototype and share demos quickly without complex frontend coding.</p><p><strong>In this episode:</strong></p><p>- Discover how Gradio’s `demo.launch(share=True)` command spins up shareable RAG UIs in minutes</p><p>- Understand environment setup challenges like nested asyncio event loops and uvloop conflicts</p><p>- Compare Gradio’s rapid prototyping advantages with production-ready custom frontends</p><p>- Learn deployment options including Hugging Face Spaces and LangChain integration</p><p>- Hear insider insights from <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>, author of “<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>”</p><p>- Discuss real-world use cases, security trade-offs, and scaling considerations</p><p><strong>Key tools &amp; technologies: </strong>Gradio, RAG pipelines, LangChain, Hugging Face Spaces, Python asyncio, nest_asyncio, uvloop</p><p><br></p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction and episode overview</p><p>02:15 - What is Gradio and why it matters for RAG</p><p>05:30 - Rapid prototyping with `demo.launch(share=True)`</p><p>08:45 - Environment quirks: asyncio loops and uvloop</p><p>11:20 - Architectural trade-offs: Gradio vs custom frontends</p><p>14:10 - Deployment strategies and hosting on Hugging Face Spaces</p><p>17:00 - Security considerations and production readiness</p><p>19:15 - Closing thoughts and resources</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for more AI engineering deep dives and practical guides</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/interfacing-rag-with-gradio-rapid-prototyping-chapter-6]]></link><guid isPermaLink="false">363e6a46-fde0-42bd-b0e8-9409d358a1ff</guid><itunes:image href="https://artwork.captivate.fm/f7c95517-f28d-4ba2-a24a-d69f0fcc8218/artwork-20251211-214520.jpg"/><pubDate>Thu, 11 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/363e6a46-fde0-42bd-b0e8-9409d358a1ff.mp3" length="23003084" type="audio/mpeg"/><itunes:duration>19:10</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>4</itunes:episode><podcast:episode>4</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/1fe38e2e-fff7-4a15-ad33-1f8e289ef341/index.html" type="text/html"/></item><item><title>Security in RAG (Chapter 5)</title><itunes:title>Security in RAG (Chapter 5)</itunes:title><description><![CDATA[<p>In this episode of Memriq Inference Digest - Engineering Edition, we explore the critical security challenges in Retrieval-Augmented Generation (RAG) systems, unpacking insights from Chapter 5 of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s '<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>.' Join us as we break down real-world vulnerabilities, defense strategies, and practical implementation patterns to build secure, production-ready RAG pipelines.</p><p><strong>In this episode:</strong></p><p>- Understand why advanced LLMs like GPT-4o can be more vulnerable to prompt probe attacks than earlier models</p><p>- Explore layered security architectures including relevance scoring and multi-LLM defenses with LangChain</p><p>- Learn how secrets management and automated adversarial testing strengthen your RAG system</p><p>- Compare manual and automated red teaming approaches and their trade-offs in production</p><p>- Hear real-world cases highlighting the legal and financial stakes of hallucinations and data leaks</p><p>- Get practical tips for building and maintaining defense-in-depth in enterprise RAG deployments</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>- OpenAI GPT-4o and GPT-3.5</p><p>- LangChain (RunnableParallel, StrOutputParser)</p><p>- python-dotenv for secrets management</p><p>- Giskard’s LLM scan for adversarial testing</p><p>- Git for version control and traceability</p><p><br></p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction and episode overview</p><p>02:30 - The surprising vulnerabilities in advanced LLMs</p><p>05:15 - Why security in RAG matters now: regulatory and technical context</p><p>07:45 - Core security concepts: retrieval as both risk and opportunity</p><p>10:30 - Comparing red teaming strategies: manual vs automated</p><p>13:00 - Under the hood: Guardian LLM architecture with LangChain</p><p>16:00 - Real-world impact: hallucinations, legal cases, and mitigation</p><p>18:30 - Practical toolbox: secrets management, relevance scoring, and continuous testing</p><p>20:00 - Closing thoughts and book spotlight</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Memriq AI: <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">https://Memriq.ai</a></p>]]></description><content:encoded><![CDATA[<p>In this episode of Memriq Inference Digest - Engineering Edition, we explore the critical security challenges in Retrieval-Augmented Generation (RAG) systems, unpacking insights from Chapter 5 of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s '<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>.' Join us as we break down real-world vulnerabilities, defense strategies, and practical implementation patterns to build secure, production-ready RAG pipelines.</p><p><strong>In this episode:</strong></p><p>- Understand why advanced LLMs like GPT-4o can be more vulnerable to prompt probe attacks than earlier models</p><p>- Explore layered security architectures including relevance scoring and multi-LLM defenses with LangChain</p><p>- Learn how secrets management and automated adversarial testing strengthen your RAG system</p><p>- Compare manual and automated red teaming approaches and their trade-offs in production</p><p>- Hear real-world cases highlighting the legal and financial stakes of hallucinations and data leaks</p><p>- Get practical tips for building and maintaining defense-in-depth in enterprise RAG deployments</p><p><strong>Key tools &amp; technologies mentioned:</strong></p><p>- OpenAI GPT-4o and GPT-3.5</p><p>- LangChain (RunnableParallel, StrOutputParser)</p><p>- python-dotenv for secrets management</p><p>- Giskard’s LLM scan for adversarial testing</p><p>- Git for version control and traceability</p><p><br></p><p><strong>Timestamps:</strong></p><p>00:00 - Introduction and episode overview</p><p>02:30 - The surprising vulnerabilities in advanced LLMs</p><p>05:15 - Why security in RAG matters now: regulatory and technical context</p><p>07:45 - Core security concepts: retrieval as both risk and opportunity</p><p>10:30 - Comparing red teaming strategies: manual vs automated</p><p>13:00 - Under the hood: Guardian LLM architecture with LangChain</p><p>16:00 - Real-world impact: hallucinations, legal cases, and mitigation</p><p>18:30 - Practical toolbox: secrets management, relevance scoring, and continuous testing</p><p>20:00 - Closing thoughts and book spotlight</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> - Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Memriq AI: <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">https://Memriq.ai</a></p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/security-in-rag-chapter-5]]></link><guid isPermaLink="false">3f6fe672-3697-4f89-aaf5-4e5a6f7472b1</guid><itunes:image href="https://artwork.captivate.fm/e4224069-622a-47ca-b37f-e32a777f81f1/artwork-20251211-201201.jpg"/><pubDate>Thu, 11 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/3f6fe672-3697-4f89-aaf5-4e5a6f7472b1.mp3" length="24886604" type="audio/mpeg"/><itunes:duration>20:44</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>3</itunes:episode><podcast:episode>3</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/80a225c4-3247-4d3d-ba0e-5a2d6c6f210b/index.html" type="text/html"/></item><item><title>RAG Components Unpacked (Chapter 4)</title><itunes:title>RAG Components Unpacked (Chapter 4)</itunes:title><description><![CDATA[<p>Unlock the engineering essentials behind Retrieval-Augmented Generation (RAG) in this episode of Memriq Inference Digest — Engineering Edition. We break down the core components of RAG pipelines as detailed in <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Chapter 4</a> of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">book</a>, exploring how offline indexing, real-time retrieval, and generation come together to solve the LLM knowledge cutoff problem.</p><p><strong>In this episode:</strong></p><p>- Explore the three-stage RAG pipeline: offline embedding and indexing, real-time retrieval, and LLM-augmented generation</p><p>- Dive into hands-on tools like LangChain, LangSmith, ChromaDB, OpenAI API, WebBaseLoader, and BeautifulSoup4</p><p>- Understand chunking strategies, embedding consistency, and pipeline orchestration with LangChain’s mini-chains</p><p>- Discuss trade-offs between direct LLM querying, offline indexing, and real-time indexing</p><p>- Hear insider insights from Keith Bourne on engineering best practices and common pitfalls</p><p>- Review real-world RAG applications in legal, healthcare, and finance domains</p><p><strong>Key tools &amp; technologies:</strong></p><p>LangChain, LangSmith, ChromaDB, OpenAI API, WebBaseLoader, BeautifulSoup4, RecursiveCharacterTextSplitter, StrOutputParser</p><p><br></p><p><strong>Timestamps:</strong></p><p>00:00 Intro &amp; overview of RAG components</p><p>03:15 The knowledge cutoff problem &amp; RAG’s architecture</p><p>06:40 Why RAG matters now: cost and tooling advances</p><p>09:10 Core RAG pipeline explained: indexing, retrieval, generation</p><p>12:00 Tool comparisons &amp; architectural trade-offs</p><p>14:30 Under the hood: code walkthrough and chunking</p><p>17:00 Real-world use cases and domain-specific insights</p><p>19:00 Final thoughts &amp; resources</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> — Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for more AI engineering guides, research breakdowns, and tools</p><p><br></p><p>Thanks for listening to Memriq Inference Digest — Engineering Edition. Stay tuned for more deep dives into AI engineering topics!</p>]]></description><content:encoded><![CDATA[<p>Unlock the engineering essentials behind Retrieval-Augmented Generation (RAG) in this episode of Memriq Inference Digest — Engineering Edition. We break down the core components of RAG pipelines as detailed in <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Chapter 4</a> of <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>’s <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">book</a>, exploring how offline indexing, real-time retrieval, and generation come together to solve the LLM knowledge cutoff problem.</p><p><strong>In this episode:</strong></p><p>- Explore the three-stage RAG pipeline: offline embedding and indexing, real-time retrieval, and LLM-augmented generation</p><p>- Dive into hands-on tools like LangChain, LangSmith, ChromaDB, OpenAI API, WebBaseLoader, and BeautifulSoup4</p><p>- Understand chunking strategies, embedding consistency, and pipeline orchestration with LangChain’s mini-chains</p><p>- Discuss trade-offs between direct LLM querying, offline indexing, and real-time indexing</p><p>- Hear insider insights from Keith Bourne on engineering best practices and common pitfalls</p><p>- Review real-world RAG applications in legal, healthcare, and finance domains</p><p><strong>Key tools &amp; technologies:</strong></p><p>LangChain, LangSmith, ChromaDB, OpenAI API, WebBaseLoader, BeautifulSoup4, RecursiveCharacterTextSplitter, StrOutputParser</p><p><br></p><p><strong>Timestamps:</strong></p><p>00:00 Intro &amp; overview of RAG components</p><p>03:15 The knowledge cutoff problem &amp; RAG’s architecture</p><p>06:40 Why RAG matters now: cost and tooling advances</p><p>09:10 Core RAG pipeline explained: indexing, retrieval, generation</p><p>12:00 Tool comparisons &amp; architectural trade-offs</p><p>14:30 Under the hood: code walkthrough and chunking</p><p>17:00 Real-world use cases and domain-specific insights</p><p>19:00 Final thoughts &amp; resources</p><p><br></p><p><strong>Resources:</strong></p><p>- "<a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Unlocking Data with Generative AI and RAG</a>" by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> — Search for 'Keith Bourne' on Amazon and grab the 2nd edition</p><p>- Visit <a href="https://Memriq.ai" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for more AI engineering guides, research breakdowns, and tools</p><p><br></p><p>Thanks for listening to Memriq Inference Digest — Engineering Edition. Stay tuned for more deep dives into AI engineering topics!</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/rag-components-unpacked-chapter-4]]></link><guid isPermaLink="false">6d8a010a-68d9-4959-8033-63386f0961a1</guid><itunes:image href="https://artwork.captivate.fm/43a09b5a-6af7-4c4f-a805-ef77181a540f/artwork-20251211-192912.jpg"/><pubDate>Thu, 11 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/6d8a010a-68d9-4959-8033-63386f0961a1.mp3" length="19665644" type="audio/mpeg"/><itunes:duration>16:23</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>2</itunes:episode><podcast:episode>2</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/a44b2b27-8185-4ebb-9a27-c7f12675bbd1/index.html" type="text/html"/></item><item><title>RAG Deep Dive: Building AI Systems That Actually Know Your Data (Chapter 1-3)</title><itunes:title>RAG Deep Dive: Building AI Systems That Actually Know Your Data (Chapter 1-3)</itunes:title><description><![CDATA[<p>In this episode, we take a deep technical dive into Retrieval-Augmented Generation (RAG), drawing heavily from <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>'s book <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank"><em>Unlocking Data with Generative AI and RAG</em></a>. We explore why RAG has become indispensable for enterprise AI systems, break down the core architecture, and share practical implementation guidance for engineers building production-grade pipelines.</p><h3>What We Cover</h3><p><em>The Problem RAG Solves</em></p><p>No matter how advanced LLMs become—GPT, Llama, Gemini, Claude—they fundamentally lack access to your private, proprietary, or real-time data. RAG bridges this gap by combining LLM reasoning with dynamic retrieval of relevant information.</p><p><em>Why RAG Is Exploding Now</em></p><ul><li>Context windows have grown dramatically (Llama 4 Scout handles up to 10M tokens)</li><li>The ecosystem has matured—LangChain alone hit 70M monthly downloads in May 2025</li><li>Infrastructure for vector storage and retrieval is production-ready</li></ul><br/><p><em>The Three-Stage Architecture</em></p><ol><li>Indexing: Convert documents into vector embeddings and store in a vector database</li><li>Retrieval: Embed user queries and perform similarity search to find relevant chunks</li><li>Generation: Feed retrieved context into an LLM prompt to generate grounded responses</li></ol><br/><p><em>RAG vs. Fine-Tuning</em></p><p>We compare trade-offs between augmenting at inference time versus modifying model weights, and discuss hybrid approaches that combine both.</p><p><em>Implementation Deep Dive</em></p><ul><li>Data ingestion and preprocessing strategies</li><li>Chunking with RecursiveCharacterTextSplitter (1,000 tokens, 200 overlap)</li><li>Embedding models and vector databases (Chroma DB, Pinecone, Weaviate)</li><li>Pipeline orchestration with LangChain Expression Language (LCEL)</li><li>Source citation patterns for compliance and auditability</li></ul><br/><p><em>Real-World Applications</em></p><p>Customer support chatbots, financial advisory systems, healthcare recommendations, ecommerce personalization, and internal knowledge bases.</p><p><em>Open Challenges</em></p><ul><li>"Lost in the middle" effect with long contexts</li><li>Multiple needles problem</li><li>Hallucination verification</li><li>Unstructured data preprocessing complexity</li></ul><br/><h3>Tools &amp; Technologies Mentioned</h3><ul><li>LangChain &amp; LlamaIndex</li><li>Chroma DB, Pinecone, Weaviate</li><li>OpenAI Embeddings</li><li>NumPy, Beautiful Soup</li><li>Meta Llama, Google Gemini, Anthropic Claude, OpenAI GPT</li></ul><br/><h3>Book Reference</h3><p><a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank"><em>Unlocking Data with Generative AI and RAG</em></a> (2nd Edition) by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> — available on Amazon. The book includes detailed diagrams, thorough explanations, and hands-on code labs for building production RAG systems.</p><p>Find <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne on LinkedIn</a>.</p><h3>Brought to You By</h3><p>Memriq — An AI content studio building practical...]]></description><content:encoded><![CDATA[<p>In this episode, we take a deep technical dive into Retrieval-Augmented Generation (RAG), drawing heavily from <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a>'s book <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank"><em>Unlocking Data with Generative AI and RAG</em></a>. We explore why RAG has become indispensable for enterprise AI systems, break down the core architecture, and share practical implementation guidance for engineers building production-grade pipelines.</p><h3>What We Cover</h3><p><em>The Problem RAG Solves</em></p><p>No matter how advanced LLMs become—GPT, Llama, Gemini, Claude—they fundamentally lack access to your private, proprietary, or real-time data. RAG bridges this gap by combining LLM reasoning with dynamic retrieval of relevant information.</p><p><em>Why RAG Is Exploding Now</em></p><ul><li>Context windows have grown dramatically (Llama 4 Scout handles up to 10M tokens)</li><li>The ecosystem has matured—LangChain alone hit 70M monthly downloads in May 2025</li><li>Infrastructure for vector storage and retrieval is production-ready</li></ul><br/><p><em>The Three-Stage Architecture</em></p><ol><li>Indexing: Convert documents into vector embeddings and store in a vector database</li><li>Retrieval: Embed user queries and perform similarity search to find relevant chunks</li><li>Generation: Feed retrieved context into an LLM prompt to generate grounded responses</li></ol><br/><p><em>RAG vs. Fine-Tuning</em></p><p>We compare trade-offs between augmenting at inference time versus modifying model weights, and discuss hybrid approaches that combine both.</p><p><em>Implementation Deep Dive</em></p><ul><li>Data ingestion and preprocessing strategies</li><li>Chunking with RecursiveCharacterTextSplitter (1,000 tokens, 200 overlap)</li><li>Embedding models and vector databases (Chroma DB, Pinecone, Weaviate)</li><li>Pipeline orchestration with LangChain Expression Language (LCEL)</li><li>Source citation patterns for compliance and auditability</li></ul><br/><p><em>Real-World Applications</em></p><p>Customer support chatbots, financial advisory systems, healthcare recommendations, ecommerce personalization, and internal knowledge bases.</p><p><em>Open Challenges</em></p><ul><li>"Lost in the middle" effect with long contexts</li><li>Multiple needles problem</li><li>Hallucination verification</li><li>Unstructured data preprocessing complexity</li></ul><br/><h3>Tools &amp; Technologies Mentioned</h3><ul><li>LangChain &amp; LlamaIndex</li><li>Chroma DB, Pinecone, Weaviate</li><li>OpenAI Embeddings</li><li>NumPy, Beautiful Soup</li><li>Meta Llama, Google Gemini, Anthropic Claude, OpenAI GPT</li></ul><br/><h3>Book Reference</h3><p><a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank"><em>Unlocking Data with Generative AI and RAG</em></a> (2nd Edition) by <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne</a> — available on Amazon. The book includes detailed diagrams, thorough explanations, and hands-on code labs for building production RAG systems.</p><p>Find <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne on LinkedIn</a>.</p><h3>Brought to You By</h3><p>Memriq — An AI content studio building practical resources for AI practitioners. Visit <a href="https://memriq.ai/" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for more engineering-focused AI content.</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/rag-deep-dive-building-ai-systems-that-actually-know-your-data-chapter-1-3]]></link><guid isPermaLink="false">fbbc9c3f-6767-4941-986d-29be3fdb3be2</guid><itunes:image href="https://artwork.captivate.fm/6207c30b-b17f-45f6-8253-801d3dc29dc0/episode-1-1-understandingRAG.jpg"/><pubDate>Thu, 11 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/fbbc9c3f-6767-4941-986d-29be3fdb3be2.mp3" length="24893804" type="audio/mpeg"/><itunes:duration>20:45</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>full</itunes:episodeType><itunes:episode>1</itunes:episode><podcast:episode>1</podcast:episode><podcast:transcript url="https://transcripts.captivate.fm/transcript/0b680fbc-3b1c-4e76-8b33-21c04290014d/index.html" type="text/html"/></item><item><title>Welcome to The Memriq AI Inference Brief</title><itunes:title>Welcome to The Memriq AI Inference Brief</itunes:title><description><![CDATA[<p>Your weekly briefing on RAG, agents, and AI memory systems.</p><p>The Memriq Inference Brief is a panel-style podcast breaking down the technologies reshaping how we build intelligent systems — from retrieval-augmented generation to agentic architectures to the emerging field of AI memory.</p><p>Two editions. Same topic. Different depths.</p><p>🎯 Leadership Edition — For executives, product leaders, and decision-makers who need to understand AI capabilities without drowning in implementation details.</p><p>🛠️ Engineering Edition — For AI engineers, data scientists, and developers who want the technical substance: architecture patterns, framework comparisons, and code you can actually use.</p><p>Topics we cover:</p><ul><li>RAG architectures and retrieval systems</li><li>Agentic AI and agent architectures</li><li>LLMs and prompt engineering</li><li>Knowledge graphs and semantic search</li><li>Vector databases and embeddings</li><li>AI memory systems (episodic, semantic, procedural)</li><li>MCP (Model Context Protocol)</li><li>Framework deep-dives and platform comparisons</li></ul><br/><p>New episodes every Monday.</p><p>A cornerstone of the content is <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne’s</a> Packt release, <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank"><em>Unlocking Data with Generative AI and RAG, 2nd edition</em></a>. We’ve built episodes and companion explanations around the concepts in the book—RAG fundamentals, GraphRAG, agent memory, intelligent recall, and how to ship this stuff responsibly.</p><p>This podcast is produced by Memriq AI — head to <a href="https://memriq.ai/" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for practical tools, guides, and deep dives built for AI practitioners.</p>]]></description><content:encoded><![CDATA[<p>Your weekly briefing on RAG, agents, and AI memory systems.</p><p>The Memriq Inference Brief is a panel-style podcast breaking down the technologies reshaping how we build intelligent systems — from retrieval-augmented generation to agentic architectures to the emerging field of AI memory.</p><p>Two editions. Same topic. Different depths.</p><p>🎯 Leadership Edition — For executives, product leaders, and decision-makers who need to understand AI capabilities without drowning in implementation details.</p><p>🛠️ Engineering Edition — For AI engineers, data scientists, and developers who want the technical substance: architecture patterns, framework comparisons, and code you can actually use.</p><p>Topics we cover:</p><ul><li>RAG architectures and retrieval systems</li><li>Agentic AI and agent architectures</li><li>LLMs and prompt engineering</li><li>Knowledge graphs and semantic search</li><li>Vector databases and embeddings</li><li>AI memory systems (episodic, semantic, procedural)</li><li>MCP (Model Context Protocol)</li><li>Framework deep-dives and platform comparisons</li></ul><br/><p>New episodes every Monday.</p><p>A cornerstone of the content is <a href="https://www.linkedin.com/in/keithbourne?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=keith_bourne_linkedin&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank">Keith Bourne’s</a> Packt release, <a href="https://a.co/d/4h3kgub?utm_source=memriq_ai&amp;utm_medium=show_notes&amp;utm_campaign=rag_book&amp;utm_content=engineering_season0_all" rel="noopener noreferrer" target="_blank"><em>Unlocking Data with Generative AI and RAG, 2nd edition</em></a>. We’ve built episodes and companion explanations around the concepts in the book—RAG fundamentals, GraphRAG, agent memory, intelligent recall, and how to ship this stuff responsibly.</p><p>This podcast is produced by Memriq AI — head to <a href="https://memriq.ai/" rel="noopener noreferrer" target="_blank">Memriq.ai</a> for practical tools, guides, and deep dives built for AI practitioners.</p>]]></content:encoded><link><![CDATA[https://memriq-inference-brief-engineering.captivate.fm/episode/welcome-to-the-memriq-inference-brief]]></link><guid isPermaLink="false">2afcc420-d358-4390-b3ce-4aa5d585d490</guid><itunes:image href="https://artwork.captivate.fm/dc49a482-f5be-4287-a390-4f8fb6151038/memriq-inference-brief.jpg"/><pubDate>Wed, 10 Dec 2025 06:00:00 -0500</pubDate><enclosure url="https://episodes.captivate.fm/episode/2afcc420-d358-4390-b3ce-4aa5d585d490.mp3" length="3168044" type="audio/mpeg"/><itunes:duration>02:38</itunes:duration><itunes:explicit>false</itunes:explicit><itunes:episodeType>trailer</itunes:episodeType><podcast:transcript url="https://transcripts.captivate.fm/transcript/9232ee69-785c-4999-8667-8412f558788f/index.html" type="text/html"/></item></channel></rss>