<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Agentic AI on Allyson Oliveira</title><link>https://www.allysonoliveira.com.br/tags/agentic-ai/</link><description>Recent content in Agentic AI on Allyson Oliveira</description><generator>Hugo -- gohugo.io</generator><language>pt-br</language><lastBuildDate>Thu, 16 Apr 2026 12:00:00 +0000</lastBuildDate><atom:link href="https://www.allysonoliveira.com.br/tags/agentic-ai/index.xml" rel="self" type="application/rss+xml"/><item><title>Playlist de AI e ML do AWS re:Invent 2025</title><link>https://www.allysonoliveira.com.br/posts/playlist-reinvent-2025-ai-ml/</link><pubDate>Thu, 16 Apr 2026 12:00:00 +0000</pubDate><guid>https://www.allysonoliveira.com.br/posts/playlist-reinvent-2025-ai-ml/</guid><description>&lt;p>O &lt;strong>AWS re:Invent&lt;/strong> é a maior conferência anual da AWS, realizada em Las Vegas. Em 2025, o evento reuniu cerca de 60.000 participantes com mais de 2.000 sessões técnicas. O tema central foi &lt;strong>IA agêntica&lt;/strong> — agentes de IA autônomos que planejam, executam e adaptam suas ações.&lt;/p>
&lt;p>Abaixo está a playlist completa de &lt;strong>Artificial Intelligence&lt;/strong> do re:Invent 2025, organizada por tema. São 67 sessões cobrindo desde infraestrutura de IA até agentes em produção.&lt;/p>
&lt;p>&lt;a class="link" href="https://www.youtube.com/playlist?list=PL2yQDdvlhXf-UqnINCmXu-dDZJm_B3bbJ" target="_blank" rel="noopener"
>Link da playlist completa no YouTube&lt;/a>&lt;/p>
&lt;h2 id="agentes-de-ia-e-amazon-bedrock-agentcore">Agentes de IA e Amazon Bedrock AgentCore
&lt;/h2>&lt;ul>
&lt;li>&lt;strong>AIM422&lt;/strong> - Agentic AI Meets Responsible AI: Strategy and best practices — Estratégias para construir agentes de IA responsáveis&lt;/li>
&lt;li>&lt;strong>AIM396&lt;/strong> - Integrate any agent framework with Amazon Bedrock AgentCore — Como integrar qualquer framework de agentes com o AgentCore&lt;/li>
&lt;li>&lt;strong>AIM3310&lt;/strong> - Agents in the enterprise: Best practices with Amazon Bedrock AgentCore — Melhores práticas para agentes em ambientes enterprise&lt;/li>
&lt;li>&lt;strong>AIM390&lt;/strong> - Building autonomous AI at scale with Amazon Bedrock — Construindo IA autônoma em escala com Bedrock&lt;/li>
&lt;li>&lt;strong>AIM3330&lt;/strong> - Keep Your Agents Out of Trouble with Amazon Bedrock AgentCore — Guardrails e segurança para agentes&lt;/li>
&lt;li>&lt;strong>AIM431&lt;/strong> - Architecting scalable and secure agentic AI with Bedrock AgentCore — Arquitetura escalável e segura para IA agêntica&lt;/li>
&lt;li>&lt;strong>AIM2204&lt;/strong> - Bridging from POC to production: An intro to Amazon Bedrock AgentCore — Do POC à produção com AgentCore&lt;/li>
&lt;li>&lt;strong>AIM3348&lt;/strong> - Improve agent quality in production with Bedrock AgentCore Evaluations — Avaliação de qualidade de agentes em produção&lt;/li>
&lt;li>&lt;strong>AIM331&lt;/strong> - Make agents remember with Amazon Bedrock AgentCore Memory — Memória persistente para agentes&lt;/li>
&lt;li>&lt;strong>AIM3313&lt;/strong> - Scale agent tools with Amazon Bedrock AgentCore Gateway — Escalando ferramentas de agentes com o Gateway&lt;/li>
&lt;li>&lt;strong>AIM395&lt;/strong> - Concept to campaign: Marketing agents on Amazon Bedrock AgentCore — Agentes de marketing com AgentCore&lt;/li>
&lt;li>&lt;strong>AIM340&lt;/strong> - AI agents for cloud ops: Automating infrastructure management — Agentes para automação de operações cloud&lt;/li>
&lt;/ul>
&lt;h2 id="amazon-nova">Amazon Nova
&lt;/h2>&lt;ul>
&lt;li>&lt;strong>AIM3324&lt;/strong> - [NEW LAUNCH] Amazon Nova 2 Omni: A new frontier in multimodal AI — Lançamento do Nova 2 Omni, modelo multimodal&lt;/li>
&lt;li>&lt;strong>AIM3342&lt;/strong> - Nova 2: Enterprise intelligence optimized for the real world — Nova 2 otimizado para cenários enterprise&lt;/li>
&lt;li>&lt;strong>AIM3325&lt;/strong> - Amazon Nova Forge: Build your own frontier models using Amazon Nova — Treine seus próprios modelos frontier com Nova Forge&lt;/li>
&lt;li>&lt;strong>AIM380&lt;/strong> - Customize Amazon Nova models for enhanced tool calling — Customização de modelos Nova para tool calling&lt;/li>
&lt;li>&lt;strong>AIM373&lt;/strong> - From prompt to production: On-brand marketing images with Amazon Nova — Geração de imagens de marketing com Nova&lt;/li>
&lt;li>&lt;strong>AIM374&lt;/strong> - Create hyper-personalized voice interactions with Amazon Nova Sonic — Interações de voz personalizadas com Nova Sonic&lt;/li>
&lt;li>&lt;strong>AIM372&lt;/strong> - Build useful, reliable agents with Amazon Nova — Construindo agentes confiáveis com Nova&lt;/li>
&lt;li>&lt;strong>AIM382&lt;/strong> - Build AI your way with Amazon Nova customization — Customização flexível do Amazon Nova&lt;/li>
&lt;li>&lt;strong>AIM384&lt;/strong> - Delighting Slack users safely and quickly with Amazon Nova and Bedrock — Slack + Nova + Bedrock&lt;/li>
&lt;li>&lt;strong>AIM429&lt;/strong> - Build autonomous code improvement agents with Amazon Nova 2 Lite — Agentes autônomos de melhoria de código com Nova 2 Lite&lt;/li>
&lt;li>&lt;strong>AIM3334&lt;/strong> - Beyond web browsers: HITL and tool integration for Nova Act — Human-in-the-loop e integração de ferramentas com Nova Act&lt;/li>
&lt;/ul>
&lt;h2 id="strands-agents-sdk">Strands Agents SDK
&lt;/h2>&lt;ul>
&lt;li>&lt;strong>AIM3331&lt;/strong> - Build production AI agents with the Strands Agents SDK for TypeScript — SDK Strands para agentes em TypeScript&lt;/li>
&lt;li>&lt;strong>AIM426&lt;/strong> - Using Strands Agents to build autonomous, self-improving AI agents — Agentes autônomos e auto-aprimoráveis com Strands&lt;/li>
&lt;li>&lt;strong>AIM3309&lt;/strong> - Slack securely powers internal AI dev tools with Bedrock and Strands — Slack + Bedrock + Strands&lt;/li>
&lt;/ul>
&lt;h2 id="amazon-sagemaker-ai">Amazon SageMaker AI
&lt;/h2>&lt;ul>
&lt;li>&lt;strong>AIM272&lt;/strong> - Master AI model development with Amazon SageMaker AI — Desenvolvimento de modelos com SageMaker&lt;/li>
&lt;li>&lt;strong>AIM363&lt;/strong> - Customize and scale foundation models using Amazon SageMaker AI — Customização e escala de foundation models&lt;/li>
&lt;li>&lt;strong>AIM364&lt;/strong> - Streamline AI model development lifecycle with Amazon SageMaker AI — Ciclo de vida de desenvolvimento de modelos&lt;/li>
&lt;li>&lt;strong>AIM424&lt;/strong> - Scaling foundation model inference on Amazon SageMaker AI — Escalando inferência de foundation models&lt;/li>
&lt;li>&lt;strong>AIM387&lt;/strong> - Scale AI agents with custom models using Amazon SageMaker AI and SGLang — Agentes com modelos customizados via SageMaker e SGLang&lt;/li>
&lt;li>&lt;strong>AIM381&lt;/strong> - Customize models for agentic AI at scale with SageMaker AI and Bedrock — Customização de modelos para IA agêntica&lt;/li>
&lt;li>&lt;strong>AIM388&lt;/strong> - Develop AI Agents faster with Amazon SageMaker Studio and Bedrock AgentCore — Desenvolvimento acelerado de agentes&lt;/li>
&lt;li>&lt;strong>AIM3340&lt;/strong> - SageMaker and MLflow: Innovate faster with no infrastructure management — SageMaker + MLflow sem gerenciar infra&lt;/li>
&lt;/ul>
&lt;h2 id="sagemaker-hyperpod-e-treinamento">SageMaker HyperPod e Treinamento
&lt;/h2>&lt;ul>
&lt;li>&lt;strong>AIM3338&lt;/strong> - SageMaker HyperPod: Checkpointless and elastic training for AI models — Treinamento elástico sem checkpoints&lt;/li>
&lt;li>&lt;strong>AIM362&lt;/strong> - Accelerate AI workloads with UltraServers on Amazon SageMaker HyperPod — UltraServers para workloads de IA&lt;/li>
&lt;li>&lt;strong>AIM371&lt;/strong> - Build, fine-tune and deploy AI models with SageMaker HyperPod CLI and SDK — CLI e SDK do HyperPod&lt;/li>
&lt;li>&lt;strong>AIM365&lt;/strong> - Train high-performing AI models at scale on AWS — Treinamento de modelos em escala&lt;/li>
&lt;li>&lt;strong>AIM3327&lt;/strong> - Unlock Advanced Model Training: Reinforcement Fine-tuning on Bedrock — Fine-tuning com reinforcement learning no Bedrock&lt;/li>
&lt;li>&lt;strong>AIM383&lt;/strong> - Build more effective agents through model customization — Agentes mais eficazes via customização de modelos&lt;/li>
&lt;/ul>
&lt;h2 id="infraestrutura-de-ia-trainium-e-nvidia">Infraestrutura de IA (Trainium e NVIDIA)
&lt;/h2>&lt;ul>
&lt;li>&lt;strong>AIM201&lt;/strong> - Break through AI performance and cost barriers with AWS Trainium — Performance e custo com Trainium&lt;/li>
&lt;li>&lt;strong>AIM3335&lt;/strong> - AWS Trn3 UltraServers: Power next-generation enterprise AI performance — Trn3 UltraServers para IA enterprise&lt;/li>
&lt;li>&lt;strong>AIM351&lt;/strong> - End-to-end foundation model lifecycle on AWS Trainium — Ciclo completo de foundation models no Trainium&lt;/li>
&lt;li>&lt;strong>AIM414&lt;/strong> - Performance engineering on Neuron: How to optimize your LLM with NKI — Otimização de LLMs com Neuron e NKI&lt;/li>
&lt;li>&lt;strong>AIM251&lt;/strong> - Accelerating AI innovation with NVIDIA GPUs on AWS — Inovação em IA com GPUs NVIDIA na AWS&lt;/li>
&lt;li>&lt;strong>AIM252&lt;/strong> - How customers build AI at scale with AWS AI infrastructure — Clientes construindo IA em escala&lt;/li>
&lt;/ul>
&lt;h2 id="rag-knowledge-bases-e-dados">RAG, Knowledge Bases e Dados
&lt;/h2>&lt;ul>
&lt;li>&lt;strong>AIM425&lt;/strong> - Advanced agentic RAG Systems: Deep dive with Amazon Bedrock — Deep dive em sistemas RAG agênticos&lt;/li>
&lt;li>&lt;strong>AIM338&lt;/strong> - Unified knowledge access: Bridging data with generative AI agents — Acesso unificado a conhecimento com agentes&lt;/li>
&lt;li>&lt;strong>AIM339&lt;/strong> - Data protection strategies for AI data foundation — Estratégias de proteção de dados para IA&lt;/li>
&lt;li>&lt;strong>AIM375&lt;/strong> - Building scalable applications with text and multimodal understanding — Aplicações escaláveis com compreensão multimodal&lt;/li>
&lt;/ul>
&lt;h2 id="amazon-bedrock-geral">Amazon Bedrock (Geral)
&lt;/h2>&lt;ul>
&lt;li>&lt;strong>AIM391&lt;/strong> - Mastering model choice: The 3-step Amazon Bedrock advantage — Como escolher o modelo certo no Bedrock&lt;/li>
&lt;li>&lt;strong>AIM3304&lt;/strong> - Balance cost, performance and reliability for AI at enterprise scale — Equilíbrio entre custo, performance e confiabilidade&lt;/li>
&lt;li>&lt;strong>AIM3341&lt;/strong> - Build Enterprise AI Apps Faster: Amazon Bedrock Multimodal Solutions — Apps enterprise com soluções multimodais do Bedrock&lt;/li>
&lt;li>&lt;strong>AIM3311&lt;/strong> - Build agentic workflows on AWS with third-party agents and tools — Workflows agênticos com agentes e ferramentas de terceiros&lt;/li>
&lt;li>&lt;strong>AIM3318&lt;/strong> - From idea to impact: Harness AI agents and tools in AWS Marketplace — Agentes e ferramentas no AWS Marketplace&lt;/li>
&lt;/ul>
&lt;h2 id="ia-responsável-e-sustentabilidade">IA Responsável e Sustentabilidade
&lt;/h2>&lt;ul>
&lt;li>&lt;strong>AIM3323&lt;/strong> - From principles to practice: Scaling AI responsibly with Indeed — IA responsável em escala com Indeed&lt;/li>
&lt;li>&lt;strong>AIM417&lt;/strong> - Sustainable computing for climate solutions — Computação sustentável para soluções climáticas&lt;/li>
&lt;li>&lt;strong>AIM253&lt;/strong> - Optimizing generative AI workloads for sustainability and cost — Otimização de workloads de IA generativa&lt;/li>
&lt;li>&lt;strong>AIM255&lt;/strong> - Architecting for sustainable IT at scale — Arquitetura sustentável em escala&lt;/li>
&lt;li>&lt;strong>AIM332&lt;/strong> - How Adobe and Salesforce enable sustainability initiatives with AWS CCFT — Adobe e Salesforce com AWS CCFT&lt;/li>
&lt;li>&lt;strong>AIM333&lt;/strong> - Sustainable and cost-efficient generative AI with agentic workflows — IA generativa sustentável com workflows agênticos&lt;/li>
&lt;li>&lt;strong>AIM237&lt;/strong> - Accelerating sustainability compliance with AI-powered document review — Compliance de sustentabilidade com IA&lt;/li>
&lt;/ul>
&lt;h2 id="casos-de-uso">Casos de Uso
&lt;/h2>&lt;ul>
&lt;li>&lt;strong>AIM256&lt;/strong> - Building an AI-powered waste classification using Amazon Nova and IoT — Classificação de resíduos com Nova e IoT&lt;/li>
&lt;li>&lt;strong>AIM336&lt;/strong> - Using AI to improve humanitarian workload resilience — IA para resiliência em workloads humanitários&lt;/li>
&lt;li>&lt;strong>AIM337&lt;/strong> - Agentic AI for member-owned financials: Systems that serve — IA agêntica para cooperativas financeiras&lt;/li>
&lt;/ul>
&lt;h2 id="well-architected-para-ia">Well-Architected para IA
&lt;/h2>&lt;ul>
&lt;li>&lt;strong>AIM (sem código)&lt;/strong> - Build a well-architected foundation for scaling generative AI and agentic apps — Fundação Well-Architected para escalar apps de IA generativa e agêntica&lt;/li>
&lt;li>&lt;strong>AIM (sem código)&lt;/strong> - Customize AI models and accelerate time to production with Amazon SageMaker AI — Customização e aceleração com SageMaker&lt;/li>
&lt;/ul>
&lt;p>Bons estudos!&lt;/p></description></item><item><title>Embeddings, Vector Database, RAG e MCP: Como os sistemas modernos de IA realmente funcionam</title><link>https://www.allysonoliveira.com.br/posts/embeddings-vectordb-rag-mcp/</link><pubDate>Thu, 16 Apr 2026 10:00:00 +0000</pubDate><guid>https://www.allysonoliveira.com.br/posts/embeddings-vectordb-rag-mcp/</guid><description>&lt;p>Recentemente assisti a um vídeo muito bom do canal &lt;a class="link" href="https://www.youtube.com/@ByteMonk" target="_blank" rel="noopener"
>ByteMonk&lt;/a> chamado &lt;strong>&amp;ldquo;Embeddings, Vector database, Agent, RAG &amp;amp; MCP: How Modern AI Systems Actually Work&amp;rdquo;&lt;/strong> (&lt;a class="link" href="https://www.youtube.com/watch?v=PByDzuOrkek" target="_blank" rel="noopener"
>link aqui&lt;/a>) e resolvi trazer um resumo dos conceitos abordados, porque são peças fundamentais para entender como os sistemas de IA modernos funcionam na prática.&lt;/p>
&lt;h2 id="embeddings-como-a-ia-entende-significado">Embeddings: como a IA entende significado
&lt;/h2>&lt;p>Embeddings são representações numéricas (vetores) que capturam o significado semântico de palavras, frases ou documentos. Em vez de tratar texto como simples sequências de caracteres, os modelos de IA transformam o conteúdo em listas de números onde &lt;strong>conceitos similares ficam próximos no espaço vetorial&lt;/strong>.&lt;/p>
&lt;p>Na prática, isso significa que uma busca por &amp;ldquo;aumentar receita&amp;rdquo; pode encontrar documentos sobre &amp;ldquo;crescimento de vendas&amp;rdquo;, mesmo sem ter palavras em comum. É isso que permite a busca semântica — ir além do match exato de palavras-chave.&lt;/p>
&lt;h2 id="vector-databases-onde-os-vetores-moram">Vector Databases: onde os vetores moram
&lt;/h2>&lt;p>Se embeddings são os vetores, as &lt;strong>Vector Databases&lt;/strong> são os sistemas de armazenamento otimizados para buscas por similaridade. Diferente de bancos de dados tradicionais que fazem buscas exatas, um vector database encontra os &amp;ldquo;vizinhos mais próximos&amp;rdquo; entre milhões de vetores em milissegundos.&lt;/p>
&lt;p>Exemplos populares: &lt;strong>Pinecone&lt;/strong>, &lt;strong>Weaviate&lt;/strong>, &lt;strong>Chroma&lt;/strong>, &lt;strong>Qdrant&lt;/strong> e &lt;strong>pgvector&lt;/strong> (extensão do PostgreSQL).&lt;/p>
&lt;p>O caso de uso clássico é armazenar chunks de documentos como vetores pesquisáveis, servindo de base para sistemas de RAG.&lt;/p>
&lt;h2 id="rag-retrieval-augmented-generation-dando-contexto-ao-llm">RAG (Retrieval-Augmented Generation): dando contexto ao LLM
&lt;/h2>&lt;p>RAG é a técnica que conecta tudo isso. O fluxo é elegante:&lt;/p>
&lt;ol>
&lt;li>O usuário faz uma pergunta&lt;/li>
&lt;li>A pergunta é convertida em um embedding&lt;/li>
&lt;li>O sistema busca documentos relevantes no vector database&lt;/li>
&lt;li>Esse contexto é injetado no prompt do LLM&lt;/li>
&lt;li>O modelo gera uma resposta baseada tanto no seu treinamento quanto nas informações recuperadas&lt;/li>
&lt;/ol>
&lt;p>O grande benefício do RAG é que o LLM consegue responder com informações &lt;strong>atualizadas e específicas do seu domínio&lt;/strong>, sem precisar ser retreinado. Isso resolve um dos maiores problemas dos modelos de linguagem: o conhecimento limitado à data de corte do treinamento.&lt;/p>
&lt;h2 id="mcp-model-context-protocol-padronizando-o-acesso-a-dados">MCP (Model Context Protocol): padronizando o acesso a dados
&lt;/h2>&lt;p>O &lt;strong>MCP&lt;/strong> (Model Context Protocol) é um protocolo que padroniza como aplicações de IA acessam fontes de dados externas. Em vez de cada integração precisar de código customizado, o MCP oferece interfaces padronizadas para que agentes de IA se conectem a bancos de dados, APIs e ferramentas de forma segura e consistente.&lt;/p>
&lt;p>Enquanto o RAG foca em recuperar informações não-estruturadas para enriquecer prompts, o MCP trata o contexto como &lt;strong>entradas dinâmicas, estruturadas e compostas&lt;/strong>, passadas via um protocolo formal. Na prática, RAG e MCP são complementares: o RAG busca o conhecimento relevante e o MCP padroniza como esse conhecimento (e outras ferramentas) chegam até o modelo.&lt;/p>
&lt;h2 id="mão-na-massa-exemplos-de-código">Mão na massa: exemplos de código
&lt;/h2>&lt;p>&lt;img src="https://www.allysonoliveira.com.br/img/rag-mcp-flow.png"
loading="lazy"
alt="Fluxo RAG com MCP"
>&lt;/p>
&lt;p>&lt;strong>Gerando um embedding com a API da OpenAI:&lt;/strong>&lt;/p>
&lt;div class="highlight">&lt;pre tabindex="0" style="color:#f8f8f2;background-color:#272822;-moz-tab-size:4;-o-tab-size:4;tab-size:4;">&lt;code class="language-python" data-lang="python">&lt;span style="display:flex;">&lt;span>&lt;span style="color:#f92672">from&lt;/span> openai &lt;span style="color:#f92672">import&lt;/span> OpenAI
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>client &lt;span style="color:#f92672">=&lt;/span> OpenAI()
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>response &lt;span style="color:#f92672">=&lt;/span> client&lt;span style="color:#f92672">.&lt;/span>embeddings&lt;span style="color:#f92672">.&lt;/span>create(
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span> model&lt;span style="color:#f92672">=&lt;/span>&lt;span style="color:#e6db74">&amp;#34;text-embedding-3-small&amp;#34;&lt;/span>,
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span> input&lt;span style="color:#f92672">=&lt;/span>&lt;span style="color:#e6db74">&amp;#34;Como funcionam os sistemas modernos de IA?&amp;#34;&lt;/span>
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>)
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>vetor &lt;span style="color:#f92672">=&lt;/span> response&lt;span style="color:#f92672">.&lt;/span>data[&lt;span style="color:#ae81ff">0&lt;/span>]&lt;span style="color:#f92672">.&lt;/span>embedding
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>print(&lt;span style="color:#e6db74">f&lt;/span>&lt;span style="color:#e6db74">&amp;#34;Dimensões do vetor: &lt;/span>&lt;span style="color:#e6db74">{&lt;/span>len(vetor)&lt;span style="color:#e6db74">}&lt;/span>&lt;span style="color:#e6db74">&amp;#34;&lt;/span>)
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>&lt;span style="color:#75715e"># Dimensões do vetor: 1536&lt;/span>
&lt;/span>&lt;/span>&lt;/code>&lt;/pre>&lt;/div>&lt;p>&lt;strong>Fazendo uma busca por similaridade com ChromaDB:&lt;/strong>&lt;/p>
&lt;div class="highlight">&lt;pre tabindex="0" style="color:#f8f8f2;background-color:#272822;-moz-tab-size:4;-o-tab-size:4;tab-size:4;">&lt;code class="language-python" data-lang="python">&lt;span style="display:flex;">&lt;span>&lt;span style="color:#f92672">import&lt;/span> chromadb
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>client &lt;span style="color:#f92672">=&lt;/span> chromadb&lt;span style="color:#f92672">.&lt;/span>Client()
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>collection &lt;span style="color:#f92672">=&lt;/span> client&lt;span style="color:#f92672">.&lt;/span>create_collection(&lt;span style="color:#e6db74">&amp;#34;meus_docs&amp;#34;&lt;/span>)
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>&lt;span style="color:#75715e"># Inserindo documentos (o Chroma gera os embeddings automaticamente)&lt;/span>
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>collection&lt;span style="color:#f92672">.&lt;/span>add(
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span> documents&lt;span style="color:#f92672">=&lt;/span>[&lt;span style="color:#e6db74">&amp;#34;RAG conecta LLMs a bases de conhecimento&amp;#34;&lt;/span>,
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span> &lt;span style="color:#e6db74">&amp;#34;Docker isola aplicações em containers&amp;#34;&lt;/span>,
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span> &lt;span style="color:#e6db74">&amp;#34;Embeddings capturam significado semântico&amp;#34;&lt;/span>],
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span> ids&lt;span style="color:#f92672">=&lt;/span>[&lt;span style="color:#e6db74">&amp;#34;doc1&amp;#34;&lt;/span>, &lt;span style="color:#e6db74">&amp;#34;doc2&amp;#34;&lt;/span>, &lt;span style="color:#e6db74">&amp;#34;doc3&amp;#34;&lt;/span>]
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>)
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>&lt;span style="color:#75715e"># Buscando por similaridade&lt;/span>
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>resultados &lt;span style="color:#f92672">=&lt;/span> collection&lt;span style="color:#f92672">.&lt;/span>query(
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span> query_texts&lt;span style="color:#f92672">=&lt;/span>[&lt;span style="color:#e6db74">&amp;#34;como dar contexto para uma IA?&amp;#34;&lt;/span>],
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span> n_results&lt;span style="color:#f92672">=&lt;/span>&lt;span style="color:#ae81ff">2&lt;/span>
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>)
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>print(resultados[&lt;span style="color:#e6db74">&amp;#34;documents&amp;#34;&lt;/span>])
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>&lt;span style="color:#75715e"># [[&amp;#39;RAG conecta LLMs a bases de conhecimento&amp;#39;,&lt;/span>
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>&lt;span style="color:#75715e"># &amp;#39;Embeddings capturam significado semântico&amp;#39;]]&lt;/span>
&lt;/span>&lt;/span>&lt;/code>&lt;/pre>&lt;/div>&lt;p>Note que a busca retornou os documentos semanticamente relevantes, mesmo sem palavras em comum com a query.&lt;/p>
&lt;h2 id="como-tudo-se-conecta">Como tudo se conecta
&lt;/h2>&lt;pre tabindex="0">&lt;code>┌──────────┐ ┌──────────────┐ ┌───────────┐ ┌─────────┐ ┌──────────┐
│ │ │ │ │ │ │ │ │ │
│ Dados ├───►│ Embeddings ├───►│ Vector DB ├───►│ RAG ├───►│ LLM │
│ │ │ │ │ │ │ │ │ │
└──────────┘ └──────────────┘ └───────────┘ └────┬────┘ └─────┬────┘
│ │
┌────▼────┐ │
│ │ │
│ MCP ◄─────────┘
│ │
└────┬────┘
│
┌─────▼──────┐
│ │
│ AI Agent │
│ │
└────────────┘
&lt;/code>&lt;/pre>&lt;p>Resumindo a stack:&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Embeddings&lt;/strong> transformam dados em vetores com significado semântico&lt;/li>
&lt;li>&lt;strong>Vector Databases&lt;/strong> armazenam e buscam esses vetores de forma eficiente&lt;/li>
&lt;li>&lt;strong>RAG&lt;/strong> usa essa infraestrutura para dar contexto relevante ao LLM em tempo de consulta&lt;/li>
&lt;li>&lt;strong>MCP&lt;/strong> padroniza a comunicação entre o agente de IA e todas essas fontes de dados e ferramentas&lt;/li>
&lt;li>&lt;strong>AI Agents&lt;/strong> orquestram tudo isso, planejando consultas em múltiplos passos e adaptando a estratégia conforme os resultados&lt;/li>
&lt;/ul>
&lt;h2 id="para-ir-além">Para ir além
&lt;/h2>&lt;ul>
&lt;li>&lt;a class="link" href="https://platform.openai.com/docs/guides/embeddings" target="_blank" rel="noopener"
>OpenAI Embeddings Guide&lt;/a> — documentação oficial sobre embeddings&lt;/li>
&lt;li>&lt;a class="link" href="https://docs.trychroma.com/" target="_blank" rel="noopener"
>ChromaDB&lt;/a> — vector database open-source, ótimo para começar&lt;/li>
&lt;li>&lt;a class="link" href="https://www.pinecone.io/learn/" target="_blank" rel="noopener"
>Pinecone Learning Center&lt;/a> — tutoriais sobre vector databases e busca semântica&lt;/li>
&lt;li>&lt;a class="link" href="https://python.langchain.com/docs/tutorials/rag/" target="_blank" rel="noopener"
>LangChain RAG Tutorial&lt;/a> — tutorial prático de RAG com Python&lt;/li>
&lt;li>&lt;a class="link" href="https://modelcontextprotocol.io/" target="_blank" rel="noopener"
>Model Context Protocol (MCP)&lt;/a> — especificação oficial do protocolo&lt;/li>
&lt;li>&lt;a class="link" href="https://www.youtube.com/watch?v=PByDzuOrkek" target="_blank" rel="noopener"
>Vídeo original do ByteMonk&lt;/a> — o vídeo que inspirou este post&lt;/li>
&lt;/ul>
&lt;p>Nos vemos por aí!&lt;/p></description></item></channel></rss>