연구
ODAR: Principled Adaptive Routing for LLM Reasoning via Active Inference
arXiv:2602.23681v1 Announce Type: new Abstract: The paradigm of large language model LLM reasoning is shifting from parameter scaling to testtime compute scaling, yet many existing approaches still rely on uniform bruteforce sampling for example, fixed bestofN or selfconsistency that is costly,...
arXiv:2602.23681v1 Announce Type: new Abstract: The paradigm of large language model LLM reasoning is shifting from parameter scaling to testtime compute scaling, yet many existing approaches still rely on uniform bruteforce sampling for example, fixed bestofN or selfconsistency that is costly, hard to attribute, and can trigger overthinking with diminishing returns.
이 콘텐츠는 ArXiv AI 원본 기사의 요약입니다. 전문은 원본 사이트에서 확인해주세요.
원문 기사 보기 →