https://babi.com/

HuggingFace 每日AI论文速递 - 2026.01.26 | LongCat练5600亿MoE代理满分;SWE-Pruner剪五成Token更快 - EarsOnMe - 精选播客,一听即合
Album
时长:
12分钟
播放:
162
发布:
1周前
主播...
简介...
https://xiaoyuzhoufm.com
【赞助商】
通勤路上就听AI每周谈。AI每周谈,每周带你回顾上周AI大事
传送门 🔗https://www.xiaoyuzhoufm.com/podcast/688a34636f5a275f1cba40fd
【目录】
本期的 15 篇论文如下:
[00:32] 🧠 LongCat-Flash-Thinking-2601 Technical Report(LongCat-Flash-Thinking-2601 技术报告)
[01:13] ✂ SWE-Pruner: Self-Adaptive Context Pruning for Coding Agents(SWE-Pruner:面向编码代理的自适应上下文剪枝框架)
[02:08] 🧠 TwinBrainVLA: Unleashing the Potential of Generalist VLMs for Embodied Tasks via Asymmetric Mixture-of-Transformers(TwinBrainVLA:通过非对称混合Transformer释放通用视觉语言模型在具身任务中的潜力)
[02:58] 🧠 VisGym: Diverse, Customizable, Scalable Environments for Multimodal Agents(VisGym:面向多模态智能体的多样化、可定制、可扩展环境)
[03:58] 🧬 Inference-Time Scaling of Verification: Self-Evolving Deep Research Agents via Test-Time Rubric-Guided Verification(验证的推理时扩展:通过测试时准则引导验证实现自演化的深度研究智能体)
[04:40] ⚡ Jet-RL: Enabling On-Policy FP8 Reinforcement Learning with Unified Training and Rollout Precision Flow(Jet-RL:通过统一的训练与推理精度流实现基于策略的FP8强化学习)
[05:32] ⚡ SALAD: Achieve High-Sparsity Attention via Efficient Linear Attention Tuning for Video Diffusion Transformer(SALAD:通过高效线性注意力调优实现视频扩散Transformer的高稀疏性注意力)
[06:11] 🧠 MeepleLM: A Virtual Playtester Simulating Diverse Subjective Experiences(MeepleLM:模拟多样化主观体验的虚拟游戏测试员)
[06:55] 🎬 Memory-V2V: Augmenting Video-to-Video Diffusion Models with Memory(Memory-V2V:利用记忆增强视频到视频扩散模型)
[07:43] 🧠 Knowledge is Not Enough: Injecting RL Skills for Continual Adaptation(知识不足够:注入强化学习技能以实现持续适应)
[08:22] 🚀 Endless Terminals: Scaling RL Environments for Terminal Agents(无尽终端:为终端智能体扩展强化学习环境)
[09:09] 🧪 DSGym: A Holistic Framework for Evaluating and Training Data Science Agents(DSGym:一个用于评估和训练数据科学智能体的整体框架)
[10:11] 🧠 Dancing in Chains: Strategic Persuasion in Academic Rebuttal via Theory of Mind(镣铐之舞:基于心智理论的学术反驳中的策略性说服)
[10:58] 💻 Guidelines to Prompt Large Language Models for Code Generation: An Empirical Characterization(面向代码生成的大语言模型提示指南:一项实证性特征研究)
[11:39] ⚖ Mecellem Models: Turkish Models Trained from Scratch and Continually Pre-trained for the Legal Domain(Mecellem模型:针对法律领域从零开始训练与持续预训练的土耳其语模型)
【关注我们】
您还可以在以下平台找到我们,获得播客内容以外更多信息
小红书: AI速递
评价...

空空如也

小宇宙热门评论...

暂无小宇宙热门评论

EarsOnMe

加入我们的 Discord

与播客爱好者一起交流

立即加入

扫描微信二维码

添加微信好友,获取更多播客资讯

微信二维码

播放列表

自动播放下一个

播放列表还是空的

去找些喜欢的节目添加进来吧