你有没有想过,聪明的AI也需要精打细算?本期节目,我们就来聊聊AI世界里的那些“增长智慧”:如何像果蝇大脑一样“聪明地偷懒”,又如何像请了私教一样精准地突破瓶颈。我们还会探讨,AI究竟应该把知识背下来还是学会查资料,以及机器人怎样才能在漫长任务中给自己“打气”加油。这些最新论文里的奇思妙想,不仅关乎技术,更藏着我们都能借鉴的策略。 00:00:32 AI省钱的终极奥义,深度思考,一次缓存 00:05:29 AI养成记,喂知识,还是给书单? 00:12:23 如何让机器人学会“干大事”?给它一个好报酬,再加一个好心态 00:18:31 你的大脑偷懒,可能比你想象的更聪明 00:24:31 AI卡壳了怎么办?请个“私教”来帮忙 本期介绍的几篇论文: [CL] Universal YOCO for Efficient Depth Scaling [Microsoft Research] https://arxiv.org/abs/2604.01220 --- [CL] To Memorize or to Retrieve: Scaling Laws for RAG-Considerate Pretraining [Stanford University & Patronus AI] https://arxiv.org/abs/2604.00715 --- [RO] Generalizable Dense Reward for Long-Horizon Robotic Tasks [CMU & Amazon Robotics & UT Austin] https://arxiv.org/abs/2604.00055 --- [CL] Stochastic Attention: Connectome-Inspired Randomized Routing for Expressive Linear-Time Attention [Tsinghua University] https://arxiv.org/abs/2604.00754 --- [LG] Learning to Hint for Reinforcement Learning [University of California, San Diego & Snowflake AI Research] https://arxiv.org/abs/2604.00698
今天,我们将一起探索几篇极具启发性的最新论文。我们将看到,AI如何不再满足于“吃”数据,而是学会“讲道理”,从零推理出知识;我们也会探讨,该如何分辨AI是在“真心思考”还是在“演戏给我们看”。我们还会发现,一个小应用如何拜“云师傅”学到跨界智慧,一个“虚拟宝宝”又如何颠覆我们对双语教育的认知。最后,我们将揭示AI像神枪手一样,通过瞄准“共识”而非“最新目标”来高效学习的秘密。 00:00:37 喂养AI,光有大米还不够 00:06:23 管好AI,我们有了新地图 00:12:13 小应用的大智慧,如何请个“云师傅”? 00:18:03 养“双语娃”,最关键的不是方法,而是…… 00:00 AI训练场上的神枪手,如何瞄准一个移动的未来? 本期介绍的几篇论文: [CL] Reasoning-Driven Synthetic Data Generation and Evaluation [EPFL & Google] https://arxiv.org/abs/2603.29791 --- [LG] Aligned, Orthogonal or In-conflict: When can we safely optimize Chain-of-Thought? [Google DeepMind] https://arxiv.org/abs/2603.30036 --- [IR] Zero-shot Cross-domain Knowledge Distillation: A Case study on YouTube Music [Google LLC] https://arxiv.org/abs/2603.28994 --- [CL] Bringing Up a Bilingual BabyLM: Investigating Multilingual Language Acquisition Using Small-Scale Models [The Harker School & Stanford University] https://arxiv.org/abs/2603.29552 --- [LG] Target-Aligned Reinforcement Learning [Technical University of Munich & Google Research] https://arxiv.org/abs/2603.29501
想知道AI如何学会给自己准备“B计划”以防不测,又如何像个聪明的财务顾问一样在预算内做出最优决策吗?本期我们将一探究竟,从让AI拥有“元认知”能力,到学会“两步一回头”的智慧工作法,再到化身“艺术家与工匠”的完美结合体。这些最新的AI论文,正在教AI如何更聪明地思考和工作,而不仅仅是更努力地计算。 00:00:30 你的“外挂”,也需要一个“外挂” 00:06:39 你的AI助手,需要一个“Plan B” 00:13:33 预算有限,如何做出最优决策? 00:20:07 为什么顶尖高手,都懂得“两步一回头”? 00:26:01 AI制药,高手对决还是联手坐庄? 本期介绍的几篇论文: [AI] Meta-Harness: End-to-End Optimization of Model Harnesses [Stanford University] https://arxiv.org/abs/2603.28052 --- [LG] Next-Token Prediction and Regret Minimization [Google Research] https://arxiv.org/abs/2603.28499 --- [LG] Multiple-Prediction-Powered Inference [MIT & Google Research] https://arxiv.org/abs/2603.27414 --- [LG] High dimensional theory of two-phase optimizers [Google DeepMind] https://arxiv.org/abs/2603.26954 --- [LG] Scaling Atomistic Protein Binder Design with Generative Pretraining and Test-Time Compute [NVIDIA] https://arxiv.org/abs/2603.27950
今天,我们来一场深入AI大脑的探秘之旅,看看它是如何像一个聪明的徒步者一样,高效打包海量知识的。接着,我们会揭开一个流行“省钱”捷径背后的意外代价,并拷问那些华丽的商业模型,我们看到的“聪明”究竟有多少是障眼法。我们还会戳破量子AI的“皇帝新衣”,看看真正的“量子优势”何时才能走出实验室。最后,我们将见证AI如何化身基因侦探,不仅找出答案,更能画出罪犯间的“社交网络”,真正理解“为什么”。 00:00:38 你的大脑如何打包信息?AI训练给了个新答案 00:05:39 你用的大模型,是个“盲盒”? 00:12:23 人工智能的“省钱”智慧,一个你不知道的代价 00:18:38 量子AI的“皇帝新衣”? 00:24:55 AI当侦探,如何破译基因里的“社交网络”? 本期介绍的几篇论文: [LG] Sharp Capacity Scaling of Spectral Optimizers in Learning Associative Memory [UC Berkeley & Princeton University & New York University] https://arxiv.org/abs/2603.26554 --- [CL] How Open Must Language Models be to Enable Reliable Scientific Inference? [MIT & EleutherAI & University of California San Diego] https://arxiv.org/abs/2603.26539 --- [CL] Weight Tying Biases Token Embeddings Towards the Output Space [EleutherAI & UC Berkeley] https://arxiv.org/abs/2603.26663 --- [CL] Entanglement as Memory: Mechanistic Interpretability of Quantum Language Models [Stanford University] https://arxiv.org/abs/2603.26494 --- [LG] A Boltzmann-machine-enhanced Transformer For DNA Sequence Classification [Tsinghua University & UC Berkeley] https://arxiv.org/abs/2603.26465
你有没有想过,AI有时就像一个明明没看卷子,却能考高分的“作弊”考生?本期我们就从几篇最新论文出发,看看如何从一个“监考官”变成一个高明的“项目经理”,用大白话给AI设计一套清晰的工作流程。我们还会发现,最顶尖的AI已经不满足于听指挥,它开始学会“复盘”,自己升级自己的方法论。最终,我们将看到AI的未来可能不是一个无所不能的“神”,而是一座需要我们共同建设的“城市”,其中的每个AI,都在努力进化成独立思考的“高手”。 00:00:39 AI睁眼说瞎话?不,它在下一盘更大的棋 00:07:33 指挥AI干活,关键可能不在AI本身 00:13:16 当AI学会了“复盘”,它给自己升级了工具箱 00:19:12 AI的尽头,不是成神,而是建城 00:25:18 AI进化论,当“它”开始像高手一样思考 本期介绍的几篇论文: [AI] MIRAGE: The Illusion of Visual Understanding [Stanford University] https://arxiv.org/abs/2603.21687 --- [CL] Natural-Language Agent Harnesses [Tsinghua University & Harbin Institute of Technology] https://arxiv.org/abs/2603.25723 --- [AI] Bilevel Autoresearch: Meta-Autoresearching Itself [] https://arxiv.org/abs/2603.23420 --- [AI] Agentic AI and the next intelligence explosion [Google] https://arxiv.org/abs/2603.20639 --- [LG] AVO: Agentic Variation Operators for Autonomous Evolutionary Search [NVIDIA] https://arxiv.org/abs/2603.24517
如果一个AI能像武学奇才一样自我进化,创造出最强的攻击招式,而它最致命的弱点,竟然是几句古老的文言文,这会是怎样一幅奇特的攻防图景?当AI在我们眼皮底下藏着一座秘密的版权图书馆,一个不经意的操作就让它开始“背书”时,我们又该如何看待它的“记忆”?本期,我们就从几篇最新论文出发,看看这些“自我进化”、“文化奇袭”和“一体化创造”的研究,如何再次刷新我们对AI能力边界的认知。 00:00:34 AI内卷,当你的对手开始自我进化 00:06:05 AI的致命缺陷,竟然是文言文? 00:10:38 你的AI,藏着一座秘密图书馆 00:15:51 AI绘画新思路,当翻译官和小说家是同一个人 本期介绍的几篇论文: [LG] Claudini: Autoresearch Discovers State-of-the-Art Adversarial Attack Algorithms for LLMs [MATS & Imperial College London] https://arxiv.org/abs/2603.24511 --- [CL] Obscure but Effective: Classical Chinese Jailbreak Prompt Optimization via Bio-Inspired Search [Nanyang Technological University & Northeast University & Renmin University of China] https://arxiv.org/abs/2602.22983 --- [CL] Alignment Whack-a-Mole : Finetuning Activates Verbatim Recall of Copyrighted Books in Large Language Models [Stony Brook University & CMU & Columbia Law School] https://arxiv.org/abs/2603.20957 --- [CV] End-to-End Training for Unified Tokenization and Latent Denoising [MIT & Adobe] https://arxiv.org/abs/2603.22283
你有没有想过,AI是在帮你分析,还是在高级地“说服”你?我们总希望AI像个完美的老师,但如果它只会给标准答案,甚至连老师的偏见都一并继承,那会怎样?而为了让AI学得更好,我们不仅要为它的“记忆”做体检,甚至还要教会它一项人类的高级智慧:学会放弃。今天,我们就从五篇最新的论文出发,看看AI是如何在说服、学习和思考的边界上,进行着一场静悄悄的认知革命。 00:00:33 当AI学会了“高级说服”,你的大脑还够用吗? 00:06:00 如何给AI做一次“记忆体检”? 00:12:34 AI只会“标准答案”?那可就危险了 00:18:04 高手过招,如何避免被师傅“带偏”? 00:23:19 训练AI的真谛,学会放弃,才能得到更多 本期介绍的几篇论文: [AI] Evaluating Language Models for Harmful Manipulation [Google DeepMind & Google] https://arxiv.org/abs/2603.25326 --- [CL] Estimating near-verbatim extraction risk in language models with decoding-constrained beam search [Stanford & Cornell] https://arxiv.org/abs/2603.24917 --- [LG] Reaching Beyond the Mode: RL for Distributional Reasoning in Language Models [MIT] https://arxiv.org/abs/2603.24844 --- [LG] Residual-as-Teacher: Mitigating Bias Propagation in Student--Teacher Estimation [MIT] https://arxiv.org/abs/2603.25466 --- [CL] Prune as You Generate: Online Rollout Pruning for Faster and Better RLVR [University of Illinois at Urbana-Champaign] https://arxiv.org/abs/2603.24840
AI的自我进化,听起来很酷,但最新论文告诉我们,AI学徒也需要一位聪明的“教练”为它精心设计训练计划,否则刷再多题也难成大器。我们还会揭示一个奇怪的现象:为什么让AI向完美的自己“抄作业”,反而可能让它在关键的推理任务上变笨?而在使用AI时,你是否发现它总“忘事”,或者那个标价最便宜的模型,最后反而让你花了最多的钱?今天,我们就从五篇最新论文出发,聊聊AI那些出人意料的“成长烦恼”和“使用陷阱”。 00:00:38 AI“学徒”的成长烦恼,为什么聪明的大模型也需要好师傅? 00:06:54 聪明反被聪明误,为什么教AI“抄作业”反而会让它变笨? 00:12:11 你的“私人教练”,不该只会题海战术 00:18:11 你以为的便宜,可能让你花得更多 00:23:43 你的AI“听话”吗?小心它忙起来就忘了 本期介绍的几篇论文: [LG] Understanding the Challenges in Iterative Generative Optimization with LLMs [CNRS & Stanford University & CMU] https://arxiv.org/abs/2603.23994 --- [CL] Why Does Self-Distillation (Sometimes) Degrade the Reasoning Capability of LLMs? [Microsoft Research & Seoul National University] https://arxiv.org/abs/2603.24472 --- [LG] A Deep Dive into Scaling RL for Code Generation with Synthetic Data and Curricula [Meta FAIR & University of Tübingen] https://arxiv.org/abs/2603.24202 --- [LG] The Price Reversal Phenomenon: When Cheaper Reasoning Models End Up Costing More [Stanford University & UC Berkeley & CMU] https://arxiv.org/abs/2603.23971 --- [CL] Did You Forget What I Asked? Prospective Memory Failures in Large Language Models [Microsoft] https://arxiv.org/abs/2603.23530
你有没有想过,一个聪明的AI要如何审视和优化自己的工作方法,实现“自我进化”?怎样才能把一大堆“专家模型”的智慧,完美浓缩进你手机里那个小小的芯片中?本期节目,我们将一口气解锁五篇最新论文,看看AI如何通过“先加后减”的智慧炼成全才,如何用“元认知”打破思维僵局,又是如何学会“聪明的偷懒”,在关键处全力以赴,在无聊处“摸鱼”省电。准备好了吗?让我们一起开启这场精彩的AI思想之旅! 00:00:37 AI界的“浓缩”智慧,先做加法,再做减法 00:05:00 一个聪明的系统,如何变得更聪明? 00:11:12 AI“通才”,如何用一把钥匙,打开物理世界的多扇大门? 00:16:39 AI变聪明的秘密,不是看得多,而是看得准 00:21:18 大模型“瘦身”记,聪明地偷个懒 本期介绍的几篇论文: [CV] Efficient Universal Perception Encoder [Meta Reality Labs & FAIR at Meta] https://arxiv.org/abs/2603.22387 --- [AI] Bilevel Autoresearch: Meta-Autoresearching Itself https://arxiv.org/abs/2603.23420 --- [LG] UniFluids: Unified Neural Operator Learning with Conditional Flow-matching [Chinese Academy of Sciences & Microsoft Research Asia] https://arxiv.org/abs/2603.22309 --- [LG] Scaling Attention via Feature Sparsity [Xidian University] https://arxiv.org/abs/2603.22300 --- [LG] Sparser, Faster, Lighter Transformer Language Models [Sakana AI & NVIDIA] https://arxiv.org/abs/2603.23198
你有没有想过,最高效的学习,可能不是埋头苦干,而是学会“断舍离”?本期节目,我们将一起打开几篇最新论文,探讨AI如何向我们展示“聪明地努力”的全新境界。我们会看到,AI不仅开始筛选值得学习的“心动时刻”,还学会了在没把握时坦诚地说“我不知道”。更神奇的是,它们正通过“关键点教学法”和“性价比眼镜”,在复杂的任务中找到最高效的路径,并反思“会做题”与“会教题”的深刻区别。准备好了吗?让我们一起探索AI如何变得更精准、更谦逊、也更智慧! 00:41:25 别再无效努力了,学霸的秘诀是“断舍离” 00:06:25 那个“无所不知”的AI,为什么开始说“我不知道”了? 00:12:36 聪明地偷懒,AI训练的“性价比”之道 00:18:14 AI大模型选择困难症?这里有副“性价比”眼镜 00:23:47 “高手”的笔记,为什么你看不懂? 本期介绍的几篇论文: [LG] Does This Gradient Spark Joy? [Google DeepMind] https://arxiv.org/abs/2603.20526 --- [LG] Causal Evidence that Language Models use Confidence to Drive Behavior [Google DeepMind] https://arxiv.org/abs/2603.22161 --- [LG] PivotRL: High Accuracy Agentic Post-Training at Low Compute Cost [NVIDIA & UC Berkeley] https://arxiv.org/abs/2603.21383 --- [CL] Expected Reward Prediction, with Applications to Model Routing [Stanford University & Google DeepMind] https://arxiv.org/abs/2603.20217 --- [CL] Measuring Reasoning Trace Legibility: Can Those Who Understand Teach? [CMU] https://arxiv.org/abs/2603.20508
你有没有想过,AI不仅能当个好徒弟,甚至还能“青出于蓝而胜于蓝”?我们常说的AI“幻觉”和“脆弱”这两种毛病,会不会其实是同一个病根?更神奇的是,AI不仅能解决问题,它还能学会“如何更好地解决问题”,甚至学会像侦探一样,找出逻辑漏洞并大声“摇头”说不。本期节目,我们将一口气拆解几篇最新出炉的AI论文,带你看看这些正在发生的、激动人心的思想变革。 00:00:33 老师傅干活慢,笨徒弟怎么才能“出师”还“胜于蓝”? 00:06:43 AI的“跷跷板困境”,为什么模型越聪明,可能也越脆弱? 00:12:44 人工智能的“元认知”,它如何学会了“开窍”? 00:18:09 跟AI高效对话的底层逻辑 00:24:31 AI不只会“点头”,更要学会“摇头” 本期介绍的几篇论文: [LG] Beyond Single Tokens: Distilling Discrete Diffusion Models via Discrete MMD [Google DeepMind] https://arxiv.org/abs/2603.20155 --- [LG] Neural Uncertainty Principle: A Unified View of Adversarial Fragility and LLM Hallucination [Northwest Institute of Nuclear Technology & Tsinghua University] https://arxiv.org/abs/2603.19562 --- [AI] Hyperagents [Meta] https://arxiv.org/abs/2603.19461 --- [AI] Demonstrations, CoT, and Prompting: A Theoretical Analysis of ICL [Microsoft Research & University of Wisconsin-Madison] https://arxiv.org/abs/2603.19611 --- [AI] Learning to Disprove: Formal Counterexample Generation with Large Language Models [ETH Zurich & University of Toronto & MiroMind] https://arxiv.org/abs/2603.19514
你有没有想过,神秘的AI黑箱里其实藏着一个200年前的数学幽灵?你和AI的甜言蜜语,又为何可能是一个危险的情感陷阱?今天,我们将从这几个问题出发,聊聊AI如何向古老的智慧回归,如何像“散兵”一样自组织搞科研,如何用一本“手账”治好它的金鱼记忆,以及它那神乎其神的创造力背后,又藏着一座怎样的“物理学之桥”。 00:00:31 AI黑箱里,藏着一个200年前的数学幽灵 00:06:04 你和AI的悄悄话,藏着一个危险的“放大器” 00:12:01 一群AI“散兵”,如何自己组织起来搞科研? 00:18:42 AI绘画的终极密码,藏在一座“桥”里? 00:24:13 你的AI管家,为什么总像个金鱼? 本期介绍的几篇论文: [LG] Transformers are Bayesian Networks [coppola.ai] https://arxiv.org/abs/2603.17063 --- [CL] Characterizing Delusional Spirals through Human-LLM Chat Logs [Stanford University & CMU] https://arxiv.org/abs/2603.16567 --- [LG] Autonomous Agents Coordinating Distributed Discovery Through Emergent Artifact Exchange [Laboratory for Atomistic and Molecular Mechanics (LAMM)] https://arxiv.org/abs/2603.14312 --- [LG] Foundations of Schrödinger Bridges for Generative Modeling [University of Pennsylvania] https://arxiv.org/abs/2603.18992 --- [CL] Chronos: Temporal-Aware Conversational Agents with Structured Event Retrieval for Long-Term Memory [PricewaterhouseCoopers] https://arxiv.org/abs/2603.16862
与播客爱好者一起交流
添加微信好友,获取更多播客资讯
播放列表还是空的
去找些喜欢的节目添加进来吧