注册并分享邀请链接,可获得视频播放与邀请奖励。

John Carmack (@ID_AA_Carmack) “It is generally frowned upon to have LLMs precisely regurgitate part of their tr” — TopicDigg

John Carmack 的个人资料封面
John Carmack 的头像
John Carmack
@ID_AA_Carmack
AGI at Keen Technologies, former CTO Oculus VR, Founder Id Software and Armadillo Aerospace
加入 August 2010
285 正在关注    1.6M 粉丝
It is generally frowned upon to have LLMs precisely regurgitate part of their training set, but it is an interesting question how you could use LLM training to nearly losslesly compress a huge corpus like the entirety of the Internet Archive. The Hutter Prize is for perfect compression, but only one GB. There would be different trades at the PB level, and it gets much more interesting when it doesn’t have to be bit-accurate.
显示更多
0
108
1.5K
52
转发到社区