Ever ask an AI a tricky question and get a hilariously wrong answer?
Nice. Does this work only with reasoning models ? Or any LLM? And have you experimented with fine tuning an LLM to implicitly do this cot ?
It works for any LLMs.
Training an LLM to do *GREAT REASONING** is very promising (e.g., deepseek R1) but also complex and costly.
Really appreciate your ability demonstrate the effectiveness of this flow with such an elegant and simple framework!
Thank you!
Nice. Does this work only with reasoning models ? Or any LLM? And have you experimented with fine tuning an LLM to implicitly do this cot ?
It works for any LLMs.
Training an LLM to do *GREAT REASONING** is very promising (e.g., deepseek R1) but also complex and costly.
Really appreciate your ability demonstrate the effectiveness of this flow with such an elegant and simple framework!
Thank you!