reStructured Pre-training 笔记
本文主要记录论文中我觉得比较重要的部分,并加入个人的理解,如有错误请可直接指出;由于格式问题,强烈建议去notion观看,完整版内容请移步notion网页进行详细阅读,谢谢!
Abstract
In such a paradigm, the role of data will be re-emphasized, and model
pre-training and fine-tuning of downstream tasks are viewed as a process of data storing and accessing.
a good storage mechanism should not only have the ability to cache a large amount of data but also consider the ease of access.
We achieve this by pre-training models over restructured data that consist of a variety of valuable information instead of raw data after overcoming several engineering challenges.
本文内容由网友自发贡献,版权归原作者所有,本站不承担相应法律责任。如您发现有涉嫌抄袭侵权的内容,请联系:hwhale#tublm.com(使用前将#替换为@)