LogoThread Easy
  • Explorer
  • Composer un thread
LogoThread Easy

Votre partenaire tout-en-un pour les threads Twitter

© 2025 Thread Easy All Rights Reserved.

Explorer

Newest first — browse tweet threads

Keep on to blur preview images; turn off to show them clearly

Thread

Thread

行道途中。非求速成,惟求通达。 2023 年扎进AI ,打通Know-How,不少赚钱项目,踩过坑,也见过光。 围城里待得够久了,出来聊聊世界,聊聊技术、聊聊赚钱。

avatar for 凡人小北
凡人小北
Sun Nov 02 16:29:56
Landing page design tip: communicate the value with a before and after visual. It converts 🔥

Landing page design tip: communicate the value with a before and after visual. It converts 🔥

UI/UX Designer ϟ Prev: lead designer at @super_ ✱ UI inspiration @DamnGoodUI

avatar for Josh Millgate
Josh Millgate
Sun Nov 02 16:29:53
could we just collectively rise as a people and just refuse to change our clocks? 4:30pm sunset is outrageous.

could we just collectively rise as a people and just refuse to change our clocks? 4:30pm sunset is outrageous.

gpus and tractors Neural networks from Scratch book: https://t.co/hyMkWyUP7R https://t.co/8WGZRkUGsn

avatar for Harrison Kinsley
Harrison Kinsley
Sun Nov 02 16:28:02
*overfit to recent data
finally, how would you create these boundaries where you decide to minimise KLD?
if temporally decided, it's suboptimal
if decided based on abstract ideas, it runs into the same challenges as BLT or LCM.

*overfit to recent data finally, how would you create these boundaries where you decide to minimise KLD? if temporally decided, it's suboptimal if decided based on abstract ideas, it runs into the same challenges as BLT or LCM.

RL and efficient distributed pretraining • eXperiments lab • memes and training lores

avatar for tokenbender
tokenbender
Sun Nov 02 16:26:15
the cost angle is definitely one of the factors. 
another is that it all gets handled at scale and we shouldn't overfit the data in pretraining that can be noisy and possibly lower quality.

the cost angle is definitely one of the factors. another is that it all gets handled at scale and we shouldn't overfit the data in pretraining that can be noisy and possibly lower quality.

*overfit to recent data finally, how would you create these boundaries where you decide to minimise KLD? if temporally decided, it's suboptimal if decided based on abstract ideas, it runs into the same challenges as BLT or LCM.

avatar for tokenbender
tokenbender
Sun Nov 02 16:24:09
https://t.co/k8dEZBW68s

https://t.co/k8dEZBW68s

Wonderer. Amor fati. Scaling trust.

avatar for Michael Frank Martin
Michael Frank Martin
Sun Nov 02 16:15:00
  • Previous
  • 1
  • More pages
  • 1310
  • 1311
  • 1312
  • More pages
  • 2117
  • Next