J A B B Y A I

Loading

how to prompt the DeepSeek-R1 model

https://preview.redd.it/niibnvu9kkhe1.png?width=680&format=png&auto=webp&s=d1fce2f1ab39e5be8293a4827fc7cbbae7861821

There’s really nothing surprising about this. Models like o1 tend to respond well to direct instructions rather than step-by-step guides or detailed chains of thought. You just have to structure the inputs clearly and use demonstrations or relevant examples to provide context instead of long explanations. I haven’t tried few-shot prompting with DeepSeek-R1 yet, but I suspect it might actually reduce o1’s performance.
My personal finds:
– Incorporating multiple languages in RL training can lead to confusing
– Geogrpahies are political driven so avoid making geographic boundaries prompt as they are highly sensitive
– Zero-shot prompt results have been great due to its Mixture of Experts.

submitted by /u/ml_guy1
[link] [comments]

Leave a Comment