LogoThread Easy
  • Explorer
  • Composer un thread
LogoThread Easy

Votre partenaire tout-en-un pour les threads Twitter

© 2025 Thread Easy All Rights Reserved.

Explorer

Newest first — browse tweet threads

Keep on to blur preview images; turn off to show them clearly

It’s always at least 12 months behind

It’s always at least 12 months behind

Code & LLMs. Working on document extraction: https://t.co/WREWHG95hE

avatar for anton
anton
Fri Nov 07 22:40:20
“Open source is now only 12 months behind”

“Open source is now only 12 months behind”

Code & LLMs. Working on document extraction: https://t.co/WREWHG95hE

avatar for anton
anton
Fri Nov 07 22:40:09
“Open source is now only 12 months behind”

“Open source is now only 12 months behind”

It’s always at least 12 months behind

avatar for anton
anton
Fri Nov 07 22:40:09
I’ve had very little success fine tuning over gpt-oss models and very much success with qwen3 models (even the instruction versions). Not sure if this is a case of skill issue or what, but they are not as friendly to tuning

I’ve had very little success fine tuning over gpt-oss models and very much success with qwen3 models (even the instruction versions). Not sure if this is a case of skill issue or what, but they are not as friendly to tuning

Code & LLMs. Working on document extraction: https://t.co/WREWHG95hE

avatar for anton
anton
Fri Oct 31 18:32:40
RT @rosinality: FP16 can have a smaller training-inference gap compared to BFloat16, thus fits better for RL. Even the difference between R…

RT @rosinality: FP16 can have a smaller training-inference gap compared to BFloat16, thus fits better for RL. Even the difference between R…

Code & LLMs. Working on document extraction: https://t.co/WREWHG95hE

avatar for anton
anton
Fri Oct 31 15:58:06
RT @eliebakouch: Training LLMs end to end is hard. Very excited to share our new blog (book?) that cover the full pipeline: pre-training, p…

RT @eliebakouch: Training LLMs end to end is hard. Very excited to share our new blog (book?) that cover the full pipeline: pre-training, p…

Code & LLMs. Working on document extraction: https://t.co/WREWHG95hE

avatar for anton
anton
Thu Oct 30 19:36:15
  • Previous
  • 1
  • 2
  • Next