Subscribe
Sign in
Share this post
arXiv Daily
2024년 5월 27일
Copy link
Facebook
Email
Notes
More
2024년 5월 27일
Kim Seonghyeon
May 27, 2024
1
Share this post
arXiv Daily
2024년 5월 27일
Copy link
Facebook
Email
Notes
More
Stacking Your Transformers: A Closer Look at Model Growth for Efficient LLM Pre-Training
Read →
Comments
Share
Copy link
Facebook
Email
Notes
More
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts
Share this post
2024년 5월 27일
Share this post
Stacking Your Transformers: A Closer Look at Model Growth for Efficient LLM Pre-Training