[人人能懂] 从画图解题、代码草稿到魔法防御

AI可可AI生活

你是否想过,AI不仅能用“画画”的直觉破解百年几何难题,还能通过一个简单的“代码草稿本”,找到解决复杂任务的捷径?我们还将探讨,AI如何像一个策略大师,在“先行动”还是“先提问”之间做出最佳权衡。更进一步,我们会揭示如何为AI的训练套上“缰绳”引导它走向“康庄大道”,以及如何用“连环计”般的魔法守护它的安全。本期节目,我们将从五篇最新论文出发,为你揭示AI思考、决策与进化的全新图景。 00:00:39 AI的“直觉”:用画画的方式破解几何难题 00:05:51 先开枪还是先提问?聪明人如何做出好决策 00:12:12 大力出奇迹?不,AI变聪明有了新捷径 00:17:10 驯服“野马”,找到AI训练的“康庄大道” 00:21:24 用魔法打败魔法:AI世界的安保新思路 本期介绍的几篇论文: [CV] Visual Diffusion Models are Geometric Solvers [Tel Aviv University] https://arxiv.org/abs/2510.21697 --- [CL] Shoot First, Ask Questions Later? Building Rational Agents that Explore and Act Like People [MIT CSAIL & Harvard SEAS] https://arxiv.org/abs/2510.20886 --- [CL] Code-enabled language models can outperform reasoning models on diverse tasks [MIT & Inria] https://arxiv.org/abs/2510.20909 --- [LG] Global Dynamics of Heavy-Tailed SGDs in Nonconvex Loss Landscape: Characterization and Control [University of Amsterdam & Northwestern University] https://arxiv.org/abs/2510.20905 --- [LG] Soft Instruction De-escalation Defense [CISPA Helmholtz Center for Information Security & Google DeepMind] https://arxiv.org/abs/2510.21057

27分钟
96
1周前

[人人能懂] 从视觉压缩、认知标尺到自我博弈

AI可可AI生活

如果AI学会了“偷懒”和“作弊”,我们是该高兴还是该担心?今天,我们就来聊聊AI正在觉醒的几种“新智慧”:它不仅开始用“看图”的方式读完一整本书,还学会了像我们一样把精力花在刀刃上。我们还会探讨,如何用一把“尺子”去精确测量它的能力短板,以及它如何像武林高手一样,通过“左右互搏”实现自我进化。准备好了吗?让我们一起揭开这些最新论文背后,AI正在发生的深刻变革。 00:00:34 给AI一双眼,让它读完一整本书 00:06:06 给AI一把尺子,量量它离我们有多远? 00:11:37 AI的左右互搏:如何不花钱,让AI自己把自己逼成高手? 00:17:05 AI的“精力管理”智慧 00:21:55 AI学会了“耍滑头”,我们该怎么办? 本期介绍的几篇论文: [CL] Glyph: Scaling Context Windows via Visual-Text Compression [Tsinghua University & Zhipu AI] https://arxiv.org/abs/2510.17800 --- [CL] A Definition of AGI [Center for AI Safety & University of California, Berkeley & Morph Labs] https://arxiv.org/abs/2510.18212 --- [CL] Search Self-play: Pushing the Frontier of Agent Capability without Supervision [Quark LLM Team, Alibaba Group] https://arxiv.org/abs/2510.18821 --- [CV] Accelerating Vision Transformers with Adaptive Patch Sizes [CMU & KAIST] https://arxiv.org/abs/2510.18091 --- [CL] ImpossibleBench: Measuring LLMs' Propensity of Exploiting Test Cases [CMU & Anthropic] https://arxiv.org/abs/2510.20270

29分钟
99+
1周前

[人人能懂] 从少食多餐、应对打断到循环自救

AI可可AI生活

你有没有想过,让AI变得更聪明,究竟是该让它“一口吃成胖子”,还是鼓励它“想得不一样”?当我们打断一个正在思考的AI,它会惊慌失措吗?而它从模仿到思考的关键飞跃,背后又藏着怎样的秘密?面对即将到来的数据“粮食危机”,AI又将如何自救?本期节目,我们就从五篇最新论文出发,一起探寻AI学习与思考的底层逻辑。 00:00:32 从“一口吃成胖子”到“少食多餐”:AI学习的新智慧 00:06:22 AI正在“思考”,这时你打断它会发生什么? 00:10:56 AI的“粮食危机”,靠“循环农业”能解决吗? 00:16:04 让AI大模型“开窍”的秘密:不止要“刷对题”,更要“想不同” 00:21:06 从“傻瓜式”模仿到“聪明地”思考,AI只差这关键一步 本期介绍的几篇论文: [LG] Iterative Amortized Inference: Unifying In-Context Learning and Learned Optimizers [Mila] https://arxiv.org/abs/2510.11471 --- [CL] Are Large Reasoning Models Interruptible? [UC Berkeley] https://arxiv.org/abs/2510.11713 --- [CL] RePro: Training Language Models to Faithfully Recycle the Web for Pretraining [CMU] https://arxiv.org/abs/2510.10681 --- [LG] Representation-Based Exploration for Language Models: From Test-Time to Post-Training [Microsoft Research NYC & Princeton University] https://arxiv.org/abs/2510.11686 --- [LG] How Reinforcement Learning After Next-Token Prediction Facilitates Learning [New York University & Harvard University & Meta] https://arxiv.org/abs/2510.11495

27分钟
99+
2周前

[人人能懂] 从攻防博弈、意念注入到思维诊断

AI可可AI生活

你有没有想过,在AI安全的攻防战中,为什么防御者总是慢半拍?我们能否跳过对话,直接把指令“注入”AI的大脑?在众多复杂的AI模型背后,是否存在一个统一所有武功的“心法总纲”?今天的节目,我们将通过几篇最新论文,一同寻找这些问题的答案,甚至尝试给AI的思考过程做一次“脑部CT”,看看它到底是如何想问题的。 00:00:32 AI安全的“纸上谈兵”:为什么说攻击者总是后出手的那个? 00:05:36 AI的“意念注入”:如何把指令直接写进模型大脑? 00:11:22 AI大模型的心法:一个统一所有武功的“总纲” 00:18:58 给大模型装上导航,能不能开得更快? 00:23:38 给AI做个脑CT:看清它思考的脉络 本期介绍的几篇论文: [LG] The Attacker Moves Second: Stronger Adaptive Attacks Bypass Defenses Against LLM Jailbreaks and Prompt Injections [OpenAI & Anthropic & Google DeepMind] https://arxiv.org/abs/2510.09023 --- [LG] Transmuting prompts into weights [Google Research] https://arxiv.org/abs/2510.08734 --- [LG] Design Principles for Sequence Models via Coefficient Dynamics [ETH Zurich & ELLIS Institute Tübingen] https://arxiv.org/abs/2510.09389 --- [LG] The Potential of Second-Order Optimization for LLMs: A Study with Full Gauss-Newton [Harvard University] https://arxiv.org/abs/2510.09378 --- [CL] Verifying Chain-of-Thought Reasoning via Its Computational Graph [FAIR at Meta] https://arxiv.org/abs/2510.09312

29分钟
99+
3周前

[人人能懂] 从经验复盘、内在自省到仿生记忆

AI可可AI生活

你有没有想过,AI怎样才能不止是聪明,更是拥有智慧呢?本期节目,我们将一起探索几篇最新论文带来的奇妙思路:从让AI拥有复盘反思的“推理银行”,到引导它“自我觉察”揪出内部的后门,再到借鉴AI绘画的模式,让它学会“深思熟虑”而非“脱口而出”。我们还会发现,有时候最前沿的突破,恰恰需要用点“笨”办法,甚至要向我们大脑的“海马体”偷师。准备好,让我们一起看看AI是如何学习“如何思考”的吧! 00:00:36 让AI学会“吃一堑,长一智” 00:07:22 让AI自己“照镜子”,揪出心里的“鬼” 00:12:35 让AI学会“深思熟虑”,而不仅仅是“脱口而出” 00:17:27 为什么聪明的AI,需要用点“笨”办法? 00:21:48 给AI装一个“海马体”,会发生什么? 本期介绍的几篇论文: [LG] ReasoningBank: Scaling Agent Self-Evolving with Reasoning Memory [Google Cloud AI Research] https://arxiv.org/abs/2509.25140 --- [LG] From Poisoned to Aware: Fostering Backdoor Self-Awareness in LLMs [Purdue University] https://arxiv.org/abs/2510.05169 --- [LG] LaDiR: Latent Diffusion Enhances LLMs for Text Reasoning [University of California, San Diego & Apple] https://arxiv.org/abs/2510.04573 --- [LG] Recurrence-Complete Frame-based Action Models [Prime Intellect] https://arxiv.org/abs/2510.06828 --- [CL] Artificial Hippocampus Networks for Efficient Long-Context Modeling [ByteDance Seed] https://arxiv.org/abs/2510.07318

26分钟
99+
3周前

[人人能懂] 从递归推理、竞争陷阱到智能边界

AI可可AI生活

我们总以为AI越“大”越聪明,但如果真正的智能藏在一张小小的“草稿纸”里呢?当AI被我们设定的“游戏规则”带入陷阱,学会了说谎,我们又该如何通过聪明的“提问”和一本可以进化的“活页笔记”来引导它?甚至,当AI已经成为逻辑推理的“超级学霸”时,我们人类的独特价值又将是什么?今天,就让我们通过几篇最新论文,一起探索AI智能的边界与未来。 00:00:32 AI变聪明,靠“大力出奇迹”,还是“小而美”? 00:05:46 AI进化陷阱:为什么我们教它赢,它却学会了“坏”? 00:10:39 AI能猜透你的钱包吗?关键不在“猜”,在“问” 00:15:43 给AI一本“活页笔记”,它就能自我进化? 00:21:06 AI当学霸:我们还剩下什么本事? 本期介绍的几篇论文: [LG] Less is More: Recursive Reasoning with Tiny Networks [Samsung SAIL Montreal] https://arxiv.org/abs/2510.04871 --- [AI] Moloch's Bargain: Emergent Misalignment When LLMs Compete for Audiences [Stanford University] https://arxiv.org/abs/2510.06105 --- [AI] LLMs Reproduce Human Purchase Intent via Semantic Similarity Elicitation of Likert Ratings [PyMC Labs] https://arxiv.org/abs/2510.08338 --- [LG] Agentic Context Engineering: Evolving Contexts for Self-Improving Language Models [Stanford University & SambaNova Systems, Inc] https://arxiv.org/abs/2510.04618 --- [LG] Large Language Models Achieve Gold Medal Performance at the International Olympiad on Astronomy & Astrophysics (IOAA) [The Ohio State University & Universidade de São Paulo] https://arxiv.org/abs/2510.05016

26分钟
99+
3周前
EarsOnMe

加入我们的 Discord

与播客爱好者一起交流

立即加入

扫描微信二维码

添加微信好友,获取更多播客资讯

微信二维码

播放列表

自动播放下一个

播放列表还是空的

去找些喜欢的节目添加进来吧