主播
节目简介
来源:小宇宙
【赞助商】
通勤路上就听AI每周谈。AI每周谈,每周带你回顾上周AI大事
传送门 🔗https://www.xiaoyuzhoufm.com/podcast/688a34636f5a275f1cba40fd
【目录】
本期的 15 篇论文如下:
[00:32] 📖 Lost in Stories: Consistency Bugs in Long Story Generation by LLMs(迷失于故事:大语言模型生成长篇故事中的一致性错误)
[01:16] 🧠 Holi-Spatial: Evolving Video Streams into Holistic 3D Spatial Intelligence(Holi-Spatial:将视频流演化为整体的3D空间智能)
[02:17] 📈 How Far Can Unsupervised RLVR Scale LLM Training?(无监督强化学习验证奖励能将LLM训练扩展到何种程度?)
[03:11] 📊 Believe Your Model: Distribution-Guided Confidence Calibration(相信你的模型:基于分布引导的置信度校准)
[04:12] 🧠 LoGeR: Long-Context Geometric Reconstruction with Hybrid Memory(LoGeR:基于混合内存的长上下文几何重建)
[05:07] 🎨 CARE-Edit: Condition-Aware Routing of Experts for Contextual Image Editing(CARE-Edit:基于条件感知专家路由的上下文图像编辑)
[05:51] 💻 CoCo: Code as CoT for Text-to-Image Preview and Rare Concept Generation(CoCo:将代码作为思维链用于文本到图像预览与稀有概念生成)
[06:30] 🎬 HiAR: Efficient Autoregressive Long Video Generation via Hierarchical Denoising(HiAR:通过分层去噪实现高效的自回归长视频生成)
[07:36] 📊 \$OneMillion-Bench: How Far are Language Agents from Human Experts?(OneMillion-Bench:语言智能体距离人类专家还有多远?)
[08:24] ⚡ NLE: Non-autoregressive LLM-based ASR by Transcript Editing(NLE:基于转录编辑的非自回归大语言模型语音识别)
[09:17] 🧠 Concept-Guided Fine-Tuning: Steering ViTs away from Spurious Correlations to Improve Robustness(概念引导的微调:引导视觉Transformer远离虚假相关性以提升鲁棒性)
[10:03] 🚀 TDM-R1: Reinforcing Few-Step Diffusion Models with Non-Differentiable Reward(TDM-R1:利用不可微奖励增强少步扩散模型)
[11:02] 📈 Unlocking Data Value in Finance: A Study on Distillation and Difficulty-Aware Training(解锁金融数据价值:关于蒸馏与难度感知训练的研究)
[11:40] 🤖 Scaling Agentic Capabilities, Not Context: Efficient Reinforcement Finetuning for Large Toolspaces(扩展智能体能力,而非上下文:面向大规模工具空间的高效强化微调)
[12:36] 🔍 PIRA-Bench: A Transition from Reactive GUI Agents to GUI-based Proactive Intent Recommendation Agents(PIRA-Bench:从反应式GUI代理到基于GUI的主动意图推荐代理的转变)
【关注我们】
您还可以在以下平台找到我们,获得播客内容以外更多信息
小红书: AI速递
通勤路上就听AI每周谈。AI每周谈,每周带你回顾上周AI大事
传送门 🔗https://www.xiaoyuzhoufm.com/podcast/688a34636f5a275f1cba40fd
【目录】
本期的 15 篇论文如下:
[00:32] 📖 Lost in Stories: Consistency Bugs in Long Story Generation by LLMs(迷失于故事:大语言模型生成长篇故事中的一致性错误)
[01:16] 🧠 Holi-Spatial: Evolving Video Streams into Holistic 3D Spatial Intelligence(Holi-Spatial:将视频流演化为整体的3D空间智能)
[02:17] 📈 How Far Can Unsupervised RLVR Scale LLM Training?(无监督强化学习验证奖励能将LLM训练扩展到何种程度?)
[03:11] 📊 Believe Your Model: Distribution-Guided Confidence Calibration(相信你的模型:基于分布引导的置信度校准)
[04:12] 🧠 LoGeR: Long-Context Geometric Reconstruction with Hybrid Memory(LoGeR:基于混合内存的长上下文几何重建)
[05:07] 🎨 CARE-Edit: Condition-Aware Routing of Experts for Contextual Image Editing(CARE-Edit:基于条件感知专家路由的上下文图像编辑)
[05:51] 💻 CoCo: Code as CoT for Text-to-Image Preview and Rare Concept Generation(CoCo:将代码作为思维链用于文本到图像预览与稀有概念生成)
[06:30] 🎬 HiAR: Efficient Autoregressive Long Video Generation via Hierarchical Denoising(HiAR:通过分层去噪实现高效的自回归长视频生成)
[07:36] 📊 \$OneMillion-Bench: How Far are Language Agents from Human Experts?(OneMillion-Bench:语言智能体距离人类专家还有多远?)
[08:24] ⚡ NLE: Non-autoregressive LLM-based ASR by Transcript Editing(NLE:基于转录编辑的非自回归大语言模型语音识别)
[09:17] 🧠 Concept-Guided Fine-Tuning: Steering ViTs away from Spurious Correlations to Improve Robustness(概念引导的微调:引导视觉Transformer远离虚假相关性以提升鲁棒性)
[10:03] 🚀 TDM-R1: Reinforcing Few-Step Diffusion Models with Non-Differentiable Reward(TDM-R1:利用不可微奖励增强少步扩散模型)
[11:02] 📈 Unlocking Data Value in Finance: A Study on Distillation and Difficulty-Aware Training(解锁金融数据价值:关于蒸馏与难度感知训练的研究)
[11:40] 🤖 Scaling Agentic Capabilities, Not Context: Efficient Reinforcement Finetuning for Large Toolspaces(扩展智能体能力,而非上下文:面向大规模工具空间的高效强化微调)
[12:36] 🔍 PIRA-Bench: A Transition from Reactive GUI Agents to GUI-based Proactive Intent Recommendation Agents(PIRA-Bench:从反应式GUI代理到基于GUI的主动意图推荐代理的转变)
【关注我们】
您还可以在以下平台找到我们,获得播客内容以外更多信息
小红书: AI速递
评价
空空如也
小宇宙热评
暂无小宇宙热门评论