Keep on to blur preview images; turn off to show them clearly

🔥 The best AI SaaS boilerplate - https://t.co/VyNtTs0jSX 🚀 The best directory boilerplate with AI - https://t.co/wEvJ1Dd8aR 🎉 https://t.co/bh1RxeERuY & https://t.co/zubXJCoY92 & https://t.co/tfQf8T7gGF


脚踏实地做梦/独立开发/降临派/新手奶爸👨🍼


3/n: DeepSeekMath-V2 model was literally threatened not to cheat. You can read it in the prompt template. Liang Wengfeng is a strict parent!!!


我的AI编程课(https://t.co/HVZn3ItASW) |B站up主 | 分享创造 + 无限迭代ing


Went from CTO to solopreneur, now building an empire one business at a time. Growing https://t.co/o8SD3o1h6p, https://t.co/NVUx34raAP, and https://t.co/2uIpQjuNAC


2/n: 'I am so dumb' moment (comparable to the 'the aha moment' of DeepSeek R1) The most important technological breakthrough in DeepSeekMath-V2 paper is not the IMO Gold level performance!!!! Then what is is? It the ability to grant model capabilities to reliably verify its own sampled generation. This has been very hard for LLMs (even the reasoning ones) Quote: "When a proof generator fails to produce a completely correct proof in one shot – common for challenging problems from competitions like IMO and CMO – iterative verification and refinement can improve results (to an extent). This involves analyzing the proof with an external verifier and prompting the generator to address identified issues. However, we observed a critical limitation: when prompted to both generate and analyze its own proof in one shot, the generator tends to claim correctness even when the external verifier easily identify flaws. In other words, while the generator can refine proofs based on external feedback, it fails to evaluate its own work with the same rigor as the dedicated verifier. This observation motivated us to endow the proof generator with genuine verification capabilities." @gm8xx8 @teortaxesTex @rohanpaul_ai @ai_for_success
