TY - GEN
T1 - Automated Peer Reviewing in Paper SEA
T2 - 2024 Findings of the Association for Computational Linguistics, EMNLP 2024
AU - Yu, Jianxiang
AU - Ding, Zichen
AU - Tan, Jiaqi
AU - Luo, Kangyang
AU - Weng, Zhenmin
AU - Gong, Chenghua
AU - Zeng, Long
AU - Cui, Renjing
AU - Han, Chengcheng
AU - Sun, Qiushi
AU - Wu, Zhiyong
AU - Lan, Yunshi
AU - Li, Xiang
N1 - Publisher Copyright:
© 2024 Association for Computational Linguistics.
PY - 2024
Y1 - 2024
N2 - In recent years, the rapid increase in scientific papers has overwhelmed traditional review mechanisms, resulting in varying quality of publications.Although existing methods have explored the capabilities of Large Language Models (LLMs) for automated scientific reviewing, their generated contents are often generic or partial.To address the issues above, we introduce an automated paper reviewing framework SEA.It comprises of three modules: Standardization, Evaluation, and Analysis, which are represented by models SEAS, SEA-E, and SEA-A, respectively.Initially, SEA-S distills data standardization capabilities of GPT-4 for integrating multiple reviews for a paper.Then, SEA-E utilizes standardized data for fine-tuning, enabling it to generate constructive reviews.Finally, SEA-A introduces a new evaluation metric called mismatch score to assess the consistency between paper contents and reviews.Moreover, we design a self-correction strategy to enhance the consistency.Extensive experimental results on datasets collected from eight venues show that SEA can generate valuable insights for authors to improve their papers.
AB - In recent years, the rapid increase in scientific papers has overwhelmed traditional review mechanisms, resulting in varying quality of publications.Although existing methods have explored the capabilities of Large Language Models (LLMs) for automated scientific reviewing, their generated contents are often generic or partial.To address the issues above, we introduce an automated paper reviewing framework SEA.It comprises of three modules: Standardization, Evaluation, and Analysis, which are represented by models SEAS, SEA-E, and SEA-A, respectively.Initially, SEA-S distills data standardization capabilities of GPT-4 for integrating multiple reviews for a paper.Then, SEA-E utilizes standardized data for fine-tuning, enabling it to generate constructive reviews.Finally, SEA-A introduces a new evaluation metric called mismatch score to assess the consistency between paper contents and reviews.Moreover, we design a self-correction strategy to enhance the consistency.Extensive experimental results on datasets collected from eight venues show that SEA can generate valuable insights for authors to improve their papers.
UR - https://www.scopus.com/pages/publications/85212691086
U2 - 10.18653/v1/2024.findings-emnlp.595
DO - 10.18653/v1/2024.findings-emnlp.595
M3 - 会议稿件
AN - SCOPUS:85212691086
T3 - EMNLP 2024 - 2024 Conference on Empirical Methods in Natural Language Processing, Findings of EMNLP 2024
SP - 10164
EP - 10184
BT - EMNLP 2024 - 2024 Conference on Empirical Methods in Natural Language Processing, Findings of EMNLP 2024
A2 - Al-Onaizan, Yaser
A2 - Bansal, Mohit
A2 - Chen, Yun-Nung
PB - Association for Computational Linguistics (ACL)
Y2 - 12 November 2024 through 16 November 2024
ER -