Loading
submitted by /u/Fabulous_Bluebird931 [link] [comments]
submitted by /u/A-Dog22 [link] [comments]
submitted by /u/sdac- [link] [comments]
submitted by /u/F0urLeafCl0ver [link] [comments]
submitted by /u/esporx [link] [comments]
submitted by /u/ml_guy1 [link] [comments]
submitted by /u/signalmutex [link] [comments]
https://www.youtube.com/watch?v=aAfanTeRn84 Lex Friedman recently posted an interview called “DeepSeek’s GPU Optimization tricks”. It is a great behind the scenes look at how Deepseek trained their latest models even when they did not have as many GPUs and their American peers.
My favorite overall benchmark is livebench. If you click show subcategories for language average you will be able to rank by plot_unscrambling which to me is the most important benchmark for writing: https://livebench.ai/ Vals is useful for tax and law
https://preview.redd.it/niibnvu9kkhe1.png?width=680&format=png&auto=webp&s=d1fce2f1ab39e5be8293a4827fc7cbbae7861821 There’s really nothing surprising about this. Models like o1 tend to respond well to direct instructions rather than step-by-step guides or detailed chains of thought. You just have to structure the inputs clearly and use demonstrations or relevant examples