时长:
6分钟
播放:
93
发布:
4个月前
主播...
简介...
[CL] Can Gradient Descent Simulate Prompting?
[MIT CSAIL]
https://arxiv.org/abs/2506.20989
---
[CL] Potemkin Understanding in Large Language Models
[MIT & University of Chicago & Harvard University]
https://arxiv.org/abs/2506.21521
---
[LG] The Ideation-Execution Gap: Execution Outcomes of LLM-Generated versus Human Research Ideas
[Stanford University]
https://arxiv.org/abs/2506.20803
---
[CL] Bridging Offline and Online Reinforcement Learning for LLMs
[FAIR at Meta]
https://arxiv.org/abs/2506.21495
---
[CL] Data Efficacy for Language Model Training
[Microsoft Research]
https://arxiv.org/abs/2506.21545
评价...
空空如也
小宇宙热门评论...
暂无小宇宙热门评论