Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I'm amazed mistral is still doing the inverse chain of thought reasoning by default, even with their new large model.

This causes it to get the question wrong for me, when testing, and only if I manually prompt normal CoT does it get it right.

Is there any papers showing a merit to this approach? It seems extremely counter-intuitive.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: