Orca 2: Enhancing Reasoning in Smaller Language Models - BigBench-Hard Subtask Metrics
2024-5-30 04:0:18 Author: hackernoon.com(查看原文) 阅读量:3 收藏

Authors:

(1) Arindam Mitra;

(2) Luciano Del Corro, work done while at Microsoft;

(3) Shweti Mahajan, work done while at Microsoft;

(4) Andres Codas, denote equal contributions;

(5) Clarisse Simoes, denote equal contributions;

(6) Sahaj Agarwal;

(7) Xuxi Chen, work done while at Microsoft;;

(8) Anastasia Razdaibiedina, work done while at Microsoft;

(9) Erik Jones, work done while at Microsoft;

(10) Kriti Aggarwal, work done while at Microsoft;

(11) Hamid Palangi;

(12) Guoqing Zheng;

(13) Corby Rosset;

(14) Hamed Khanpour;

(15) Ahmed Awadall.

Table 7, 8, 9, and 10 showcase the zero-shot performance of Orca 2 and the baseline models on each BBH MCQ reasoning task, with accuracy being the metric used to evaluate performance.


文章来源: https://hackernoon.com/orca-2-enhancing-reasoning-in-smaller-language-models-bigbench-hard-subtask-metrics?source=rss
如有侵权请联系:admin#unsafe.sh