主播
节目简介
来源:小宇宙
【赞助商】
通勤路上就听AI每周谈。AI每周谈,每周带你回顾上周AI大事
传送门 🔗https://www.xiaoyuzhoufm.com/podcast/688a34636f5a275f1cba40fd
【目录】
本期的 15 篇论文如下:
[00:29] 🧠 Self-Distilled RLVR(基于自蒸馏的强化学习与可验证奖励)
[01:18] 🎯 A Simple Baseline for Streaming Video Understanding(流式视频理解的简单基线)
[02:07] 🔍 Token Warping Helps MLLMs Look from Nearby Viewpoints(Token扭曲助力多模态大语言模型从邻近视角观察)
[03:06] 🔍 Agentic-MME: What Agentic Capability Really Brings to Multimodal Intelligence?(Agentic-MME:能动性能力究竟为多模态智能带来了什么?)
[03:57] 📈 Test-Time Scaling Makes Overtraining Compute-Optimal(测试时扩展使过度训练达到计算最优)
[04:56] 🧠 Communicating about Space: Language-Mediated Spatial Integration Across Partial Views(空间交流:跨局部视角的语言中介空间整合)
[05:39] 🏆 GrandCode: Achieving Grandmaster Level in Competitive Programming via Agentic Reinforcement Learning(GrandCode:通过智能体强化学习在竞技编程中达到宗师级水平)
[06:27] 🤖 InCoder-32B-Thinking: Industrial Code World Model for Thinking(InCoder-32B-Thinking:面向思考的工业代码世界模型)
[07:22] 🛡 AgentSocialBench: Evaluating Privacy Risks in Human-Centered Agentic Social Networks(AgentSocialBench:评估以人为中心的代理社交网络中的隐私风险)
[08:10] ⚠ AgentHazard: A Benchmark for Evaluating Harmful Behavior in Computer-Use Agents(AgentHazard:计算机使用智能体有害行为评估基准)
[08:52] ⚡ Swift-SVD: Theoretical Optimality Meets Practical Efficiency in Low-Rank LLM Compression(Swift-SVD:理论最优性与实际效率在低秩大语言模型压缩中的结合)
[09:39] 🔍 VLMs Need Words: Vision Language Models Ignore Visual Detail In Favor of Semantic Anchors(视觉语言模型需要词汇:视觉语言模型忽略视觉细节而依赖语义锚点)
[10:30] 📊 Xpertbench: Expert Level Tasks with Rubrics-Based Evaluation(Xpertbench:基于量规评估的专家级任务基准)
[11:16] 🎬 Salt: Self-Consistent Distribution Matching with Cache-Aware Training for Fast Video Generation(Salt:用于快速视频生成的自洽分布匹配与缓存感知训练)
[12:04] 🤝 CoME-VL: Scaling Complementary Multi-Encoder Vision-Language Learning(CoME-VL:扩展互补多编码器视觉语言学习)
【关注我们】
您还可以在以下平台找到我们,获得播客内容以外更多信息
小红书: AI速递
通勤路上就听AI每周谈。AI每周谈,每周带你回顾上周AI大事
传送门 🔗https://www.xiaoyuzhoufm.com/podcast/688a34636f5a275f1cba40fd
【目录】
本期的 15 篇论文如下:
[00:29] 🧠 Self-Distilled RLVR(基于自蒸馏的强化学习与可验证奖励)
[01:18] 🎯 A Simple Baseline for Streaming Video Understanding(流式视频理解的简单基线)
[02:07] 🔍 Token Warping Helps MLLMs Look from Nearby Viewpoints(Token扭曲助力多模态大语言模型从邻近视角观察)
[03:06] 🔍 Agentic-MME: What Agentic Capability Really Brings to Multimodal Intelligence?(Agentic-MME:能动性能力究竟为多模态智能带来了什么?)
[03:57] 📈 Test-Time Scaling Makes Overtraining Compute-Optimal(测试时扩展使过度训练达到计算最优)
[04:56] 🧠 Communicating about Space: Language-Mediated Spatial Integration Across Partial Views(空间交流:跨局部视角的语言中介空间整合)
[05:39] 🏆 GrandCode: Achieving Grandmaster Level in Competitive Programming via Agentic Reinforcement Learning(GrandCode:通过智能体强化学习在竞技编程中达到宗师级水平)
[06:27] 🤖 InCoder-32B-Thinking: Industrial Code World Model for Thinking(InCoder-32B-Thinking:面向思考的工业代码世界模型)
[07:22] 🛡 AgentSocialBench: Evaluating Privacy Risks in Human-Centered Agentic Social Networks(AgentSocialBench:评估以人为中心的代理社交网络中的隐私风险)
[08:10] ⚠ AgentHazard: A Benchmark for Evaluating Harmful Behavior in Computer-Use Agents(AgentHazard:计算机使用智能体有害行为评估基准)
[08:52] ⚡ Swift-SVD: Theoretical Optimality Meets Practical Efficiency in Low-Rank LLM Compression(Swift-SVD:理论最优性与实际效率在低秩大语言模型压缩中的结合)
[09:39] 🔍 VLMs Need Words: Vision Language Models Ignore Visual Detail In Favor of Semantic Anchors(视觉语言模型需要词汇:视觉语言模型忽略视觉细节而依赖语义锚点)
[10:30] 📊 Xpertbench: Expert Level Tasks with Rubrics-Based Evaluation(Xpertbench:基于量规评估的专家级任务基准)
[11:16] 🎬 Salt: Self-Consistent Distribution Matching with Cache-Aware Training for Fast Video Generation(Salt:用于快速视频生成的自洽分布匹配与缓存感知训练)
[12:04] 🤝 CoME-VL: Scaling Complementary Multi-Encoder Vision-Language Learning(CoME-VL:扩展互补多编码器视觉语言学习)
【关注我们】
您还可以在以下平台找到我们,获得播客内容以外更多信息
小红书: AI速递