LogoThread Easy
  • 発見
  • スレッド作成
LogoThread Easy

Twitter スレッドの万能パートナー

© 2025 Thread Easy All Rights Reserved.

探索

Newest first — browse tweet threads

Keep on to blur preview images; turn off to show them clearly

Heat-related deaths have surged 23% since the 1990s, reaching 546,000 annually, as climate change drives rising global temperatures

Heat-related deaths have surged 23% since the 1990s, reaching 546,000 annually, as climate change drives rising global temperatures

Top and breaking news, pictures and videos from Reuters. For breaking business news, follow @ReutersBiz. Our daily podcast is here: https://t.co/KO0QFy0d3a

avatar for Reuters
Reuters
Wed Oct 29 18:40:00
RT @neuranne: Tiny Experiments sold over 60K copies since March!! 🥹🙏❤️‍🔥🙌🥳 so grateful to everyone who bought the book and helped spread th…

RT @neuranne: Tiny Experiments sold over 60K copies since March!! 🥹🙏❤️‍🔥🙌🥳 so grateful to everyone who bought the book and helped spread th…

hypercurious :) founder @ness_labs • neuroscientist @KingsIoPPN • author of Tiny Experiments • personal science, systematic curiosity, experimental thinking ꩜⋆✦

avatar for Anne-Laure Le Cunff
Anne-Laure Le Cunff
Wed Oct 29 18:39:55
I just realised that the day we were attacked in ABJ was literally one year exactly to the date that this video was uploaded

I just realised that the day we were attacked in ABJ was literally one year exactly to the date that this video was uploaded

Founder | Author | Speaker Building @beltstripe I'm Not The Man Of Your Dreams. Your Imagination Wasn't This Great.

avatar for Sani Yusuf
Sani Yusuf
Wed Oct 29 18:38:15
The thing I'm doing on today with GPT-5 Pro is interesting because it's exactly the thing I was told was impossible: a deep, multi-step pair refactoring that pays down tech debt (somewhat accrued due to aggressive AI use) and towards re-using code due to totally new requirements

The thing I'm doing on today with GPT-5 Pro is interesting because it's exactly the thing I was told was impossible: a deep, multi-step pair refactoring that pays down tech debt (somewhat accrued due to aggressive AI use) and towards re-using code due to totally new requirements

Opening portals to handheld VR at https://t.co/A2JMItorCV. Problems soluble, potential to improve invariant.

avatar for gfodor.id
gfodor.id
Wed Oct 29 18:38:10
4/4 Physical Infrastructure for Cursor's Composer model training.

They claim to have trained (and continue to train) across thousands of GPUs. They train models in low precision, and they use asynchronous RL (next tweet to explain what it is).

Quote: "We built custom training infrastructure leveraging PyTorch and Ray to power asynchronous reinforcement learning at scale. 

We natively train our models at low precision by combining our MXFP8 MoE kernels with expert parallelism and hybrid sharded data parallelism, allowing us to scale training to thousands of NVIDIA GPUs with minimal communication cost. 

Additionally, training with MXFP8 allows us to deliver faster inference speeds without requiring post-training quantisation."

4/4 Physical Infrastructure for Cursor's Composer model training. They claim to have trained (and continue to train) across thousands of GPUs. They train models in low precision, and they use asynchronous RL (next tweet to explain what it is). Quote: "We built custom training infrastructure leveraging PyTorch and Ray to power asynchronous reinforcement learning at scale. We natively train our models at low precision by combining our MXFP8 MoE kernels with expert parallelism and hybrid sharded data parallelism, allowing us to scale training to thousands of NVIDIA GPUs with minimal communication cost. Additionally, training with MXFP8 allows us to deliver faster inference speeds without requiring post-training quantisation."

5/5 What is async RL that Customer Composer model training uses? It uses asynchronous execution at multiple levels to avoid waiting on slow operations e.g. a long roll-out generation. As you know, for a given problem, in RL like GRPO we generate multiple trajectorier. However, some trajectories can take too long to complete. So, once they have enough trajectories, they run the training. Partial samples/roll-outs are resumed later with updated model. This causes a situation where some tokens are generated by the old model/policy and some by new. However, this is acceptable. If you want to understand more about Async RL, please read APRIL - a project for Async RL.

avatar for GDP
GDP
Wed Oct 29 18:37:19
i have a bunch of markov chain monte carlo stuff lying around from grad school i can throw at this, just have to port it from matlab

i have a bunch of markov chain monte carlo stuff lying around from grad school i can throw at this, just have to port it from matlab

back in the prehistory of 3d computer vision (2016) we would use probabilistic / ebm models to fit shape templates to street scenes

avatar for anton 🇺🇸
anton 🇺🇸
Wed Oct 29 18:35:53
  • Previous
  • 1
  • More pages
  • 1913
  • 1914
  • 1915
  • More pages
  • 2118
  • Next