Zetaphor@zemmy.cc to LocalLLaMA@sh.itjust.worksEnglish · 2 years agoDistilling step-by-step: Outperforming larger language models with less training data and smaller model sizesblog.research.googleexternal-linkmessage-square3fedilinkarrow-up128cross-posted to: generative_ai@mander.xyzhackernews@lemmy.smeargle.fanshackernews@derp.foo
arrow-up128external-linkDistilling step-by-step: Outperforming larger language models with less training data and smaller model sizesblog.research.googleZetaphor@zemmy.cc to LocalLLaMA@sh.itjust.worksEnglish · 2 years agomessage-square3fedilinkcross-posted to: generative_ai@mander.xyzhackernews@lemmy.smeargle.fanshackernews@derp.foo
minus-squareZetaphor@zemmy.ccOPlinkfedilinkEnglisharrow-up2·2 years agoThe code is available here: https://github.com/google-research/distilling-step-by-step
minus-squarenoneabove1182@sh.itjust.worksMlinkfedilinkEnglisharrow-up1·2 years agoSomehow this is even more confusing because that code hasn’t been touched in 3 months, maybe just took them that long to validate? Will have to read through it, thanks!
The code is available here:
https://github.com/google-research/distilling-step-by-step
Somehow this is even more confusing because that code hasn’t been touched in 3 months, maybe just took them that long to validate? Will have to read through it, thanks!