Powered by RND

AI Odyssey

Anlie Arnaudy, Daniel Herbera and Guillaume Fournier
AI Odyssey
Dernier épisode

Épisodes disponibles

5 sur 53
  • Will Your Next Prompt Engineer Be an AI?
     What if you could get the performance of a massive, 100-example prompt, but with 13 times fewer tokens?That’s the breakthrough promise of "instruction induction" —teaching an AI to be the prompt engineer.This week, we dive into PROMPT-MII , a new framework that essentially meta-learns how to write compact, high-performance instructions for LLMs. It’s a reinforcement learning approach that could make AI adaptation both cheaper and more effective.This episode explores the original research by Emily Xiao, Yixiao Zeng, Ada Chen, Chin-Jou Li, Amanda Bertsch, and Graham Neubig from Carnegie Mellon University.Read the full paper here for a deeperdive: https://arxiv.org/abs/2510.16932
    --------  
    17:58
  • The Vision Hack: How a Picture Solved AI's Biggest Memory Problem
    The biggest bottleneck for AIs handling massive documents—the context window—just got a radical fix. DeepSeek AI's DeepSeek-GOCR uses a counterintuitive trick: it turns text into an image to compress it by up to 10 times without losing accuracy. That means your AI can suddenly read the equivalent of 20 million tokens (entire codebases or legal troves) efficiently! This episode dives into the elegant vision-based solution, the power of its Mixture of Experts architecture, and why some experts believe all AI input should become an image.Original Research: DeepSeek-GOCR is a breakthrough by the DeepSeek AI team.Content generated with the help of Google's NotebookLM.Link to the Original Research Paper: https://deepseek.ai/blog/deepseek-ocr-context-compression
    --------  
    14:22
  • Smarter Agents, Less Budget: Reinforcement Learning with Tree Search
    Training AI agents using Reinforcement Learning (RL) to handle complex, multi-turn tasks is notoriously difficult.Traditional methods face two major hurdles: high computational costs (generating numerous interaction scenarios, or "rollouts," is expensive) and sparse supervision (rewards are only given at the very end of a task, making it hard for the agent to learn which specific steps were useful).In this episode, we explore "Tree Search for LLM Agent Reinforcement Learning," by researchers from Xiamen University, AMAP (Alibaba Group), and the Southern University of Science and Technology. They introduce a novel approach called Tree-GRPO (Tree-based Group Relative Policy Optimization) that fundamentally changes how agents explore possibilities.Tree-GRPO replaces inefficient "chain-based" sampling with a tree-search strategy. By allowing different trajectories to share common prefixes (the initial steps of an interaction), the method significantly increases the number of scenarios explored within the same budget. Crucially, the tree structure allows the system to derive step-by-step "process supervision signals," even when only the final outcome reward is available. The results demonstrate superior performance over traditional methods, with some models achieving better results using only a quarter of the training budget.📄 Paper: Tree Search for LLM Agent Reinforcement Learning https://arxiv.org/abs/2509.21240
    --------  
    0:35
  • Beyond the AI Agent Builders Hype
    Everyone's talking about AI agents that can automate complex tasks. But what happens when a cool demo meets the real world? We dive into hard-won, and often surprising, lessons from builders on the front lines. Discover why your first strategic choice isn't about a tool, but an entire ecosystem; why more agents can actually make things worse; and why the most critical skill is shifting from "prompt engineering" to "context engineering." This episode cuts through the noise to reveal what it really takes to build reliable AI agents that deliver value.
    --------  
    14:07
  • AI That Quietly Helps: Overhearing Agents
    In this IA Odyssey episode, we unpack “overhearing agents”—AI systems that listen to human activity (audio, text, or video) and step in only when help is useful, like surfacing a diagram during a class discussion, prepping trail options while a family plans a hike, or pulling case notes in a medical consult.While conversational AI (like chatbots) requires direct user engagement, overhearing agents continuously monitor ambient activities, such as human-to-human conversations, and intervene only to provide contextual assistance without interruption. Examples include silently providing data during a medical consultation or scheduling meetings as colleagues discuss availability.The paper introduces a clear taxonomy for how these agents activate: always-on, user-initiated, post-hoc analysis, or rule-based triggers. This framework helps developers think about when and how an AI should “step in” without becoming intrusive.Original paper: https://arxiv.org/pdf/2509.16325Credits: Episode notes synthesized with Google’s NotebookLM to analyze and summarize the paper; all insights credit the original authors.
    --------  
    0:43

Plus de podcasts Technologies

À propos de AI Odyssey

AI Odyssey is your journey through the vast and evolving world of artificial intelligence. Powered by AI, this podcast breaks down both the foundational concepts and the cutting-edge developments in the field. Whether you're just starting to explore the role of AI in our world or you're a seasoned expert looking for deeper insights, AI Odyssey offers something for everyone. From AI ethics to machine learning intricacies, each episode is crafted to inspire curiosity and spark discussion on how artificial intelligence is shaping our future.
Site web du podcast

Écoutez AI Odyssey, Monde Numérique (Actu Tech) ou d'autres podcasts du monde entier - avec l'app de radio.fr

Obtenez l’app radio.fr
 gratuite

  • Ajout de radios et podcasts en favoris
  • Diffusion via Wi-Fi ou Bluetooth
  • Carplay & Android Auto compatibles
  • Et encore plus de fonctionnalités
Applications
Réseaux sociaux
v7.23.11 | © 2007-2025 radio.de GmbH
Generated: 11/7/2025 - 9:32:53 PM