HyperAIHyperAI

Command Palette

Search for a command to run...

L'IA peut acquérir le goût scientifique.

Résumé

Les grands scientifiques possèdent un jugement aiguisé et une vision prospective, étroitement liés à ce que l'on nomme le « goût scientifique ». Dans ce travail, nous utilisons ce terme pour désigner la capacité à évaluer et à proposer des idées de recherche à fort potentiel d'impact. Toutefois, la plupart des études connexes se concentrent sur l'amélioration des capacités exécutives d'un scientifique artificiel, tandis que le renforcement du goût scientifique des systèmes d'IA demeure largement inexploré. Dans cette étude, nous proposons l'apprentissage par renforcement à partir de retours communautaires (Reinforcement Learning from Community Feedback, RLCF), un paradigme d'entraînement exploitant des signaux communautaires à grande échelle comme supervision, et nous formulons l'apprentissage du goût scientifique comme un problème de modélisation et d'alignement des préférences. Pour la modélisation des préférences, nous entraînons un modèle évaluateur, Scientific Judge, sur 700 000 paires d'articles à haute et faible citation, appariées selon le domaine et l'année, afin d'évaluer les idées de recherche. Pour l'alignement des préférences, en utilisant Scientific Judge comme modèle de récompense, nous entraînons un modèle de politique, Scientific Thinker, afin qu'il génère des idées de recherche à fort potentiel d'impact. Les résultats expérimentaux montrent que Scientific Judge surpasse les LLMs de l'état de l'art (par exemple, GPT-5.2, Gemini 3 Pro) et généralise efficacement à des tests portant sur des années futures, des domaines inédits ainsi qu'aux préférences issues de l'évaluation par les pairs. De plus, Scientific Thinker propose des idées de recherche présentant un potentiel d'impact supérieur à celui des modèles de référence. Ces résultats démontrent que l'IA peut acquérir un goût scientifique, marquant une étape clé vers la réalisation de scientifiques artificiels au niveau humain.

One-sentence Summary

Researchers from Fudan University and collaborating institutes propose RLCF, a training paradigm that leverages community feedback to enhance AI scientific taste. By deploying the SCIENTIFIC JUDGE and SCIENTIFIC THINKER models, this work shifts focus from executive capability to generating high-impact research ideas, outperforming state-of-the-art systems in predicting future scientific success.

Key Contributions

  • The paper introduces Reinforcement Learning from Community Feedback (RLCF), a training paradigm that frames scientific taste learning as a preference modeling and alignment problem using large-scale community signals like citations as supervision.
  • A new dataset called SciJUDGEBench is constructed containing 700K field- and time-matched pairs of high- versus low-citation paper abstracts to enable the training and evaluation of AI scientific judgment.
  • Two specialized models are developed where SCIENTIFIC JUDGE outperforms state-of-the-art LLMs in predicting paper impact across unseen fields and future years, while SCIENTIFIC THINKER generates research ideas with demonstrably higher potential impact than baseline methods.

Introduction

Current efforts to build AI scientists primarily focus on executive capabilities like literature search and automated experimentation, yet they struggle to replicate the human ability to judge and propose high-impact research directions, a skill known as scientific taste. Existing approaches often rely on costly human annotations or fail to capture the broader community consensus that drives scientific progress, leaving a gap in AI's intrinsic judgment and ideation potential. To address this, the authors introduce Reinforcement Learning from Community Feedback (RLCF), a paradigm that treats scientific taste as a preference modeling problem trained on large-scale citation signals. They construct SciJUDGEBench with 700K matched paper pairs to train SCIENTIFIC JUDGE, a reward model that outperforms state-of-the-art LLMs in predicting impact, and subsequently use it to train SCIENTIFIC THINKER, a policy model that generates research ideas with significantly higher potential impact than baseline methods.

Dataset

  • Dataset Composition and Sources The authors construct SciJUDGEBENCH from 2.1 million arXiv papers published through 2024, covering Computer Science, Mathematics, Physics, and a diverse "Others" category that includes Economics, Quantitative Biology, and Statistics. The dataset transforms community feedback into pairwise supervision signals by pairing scientific ideas represented by their titles and abstracts.

  • Key Details for Each Subset

    • Training Set: Contains 696,758 field- and time-matched preference pairs derived from 1.4 million unique papers. Pairs are filtered to ensure the preferred paper has an absolute citation difference of at least 8 and a relative difference of at least 30% compared to the lower-cited paper.
    • Main Test Set: Comprises 728 in-domain pairs with stricter filtering requiring an absolute citation difference greater than 32 and a relative difference of at least 50% to ensure clear preference signals.
    • Temporal OOD Test Set: Includes 514 pairs from papers published in 2025 to test extrapolation to future data. This set uses adaptive percentile-based thresholds within subcategories to pair high-citation papers with low-citation contemporaries while maintaining a 5-day publication window.
    • Metric OOD Test Set: Consists of 611 pairs from ICLR submissions (2017–2026) where preferences are determined by peer review scores rather than citations. The authors filter out papers with low reviewer confidence or high rating variance, then pair top and bottom performers to test transferability to review-based judgment.
    • BioRxiv Test Set: A smaller set of 160 biology pairs is also reported for additional evaluation.
  • Model Training and Usage The authors train the SCIENTIFIC JUDGE model on various base architectures, including the Qwen2.5-Instruct series (1.5B to 32B parameters), Qwen3 variants, and Llama-3.1-8B-Instruct. They employ Group Relative Policy Optimization (GRPO) using preference prediction correctness as a verifiable reward. The model generates a reasoning trace followed by a binary prediction (A or B) and receives a reward of 1 only if the prediction is correct.

  • Processing and Evaluation Strategies To mitigate field and time biases, all pairs are matched within the same subcategory and similar publication time windows. The authors address position bias during evaluation by presenting each pair twice with swapped orders and scoring a prediction as correct only if the model remains consistent across both orderings. Metadata construction relies on primary arXiv categories, and the "Others" field is explicitly aggregated from specific subcategories rather than treated as a residual bucket.

Method

The authors propose Reinforcement Learning from Community Feedback (RLCF), a three-stage training paradigm designed to instill scientific taste in large language models. This framework leverages large-scale community signals to supervise the learning process, progressing from data construction to preference modeling and finally to preference alignment.

The process begins with the collection of community feedback to construct preference data. A corpus of scientific papers is gathered, and pairs are matched based on their field and publication time to ensure fair comparison. Citations serve as natural preference signals, where the paper with higher citation counts is deemed to have greater potential impact. Formally, the cumulative expected impact of a paper ppp is defined as I(p)=limNt=1NE[ct(p)]I(p) = \lim_{N \to \infty} \sum_{t=1}^N \mathbb{E}[c_t(p)]I(p)=limNt=1NE[ct(p)], where ct(p)c_t(p)ct(p) represents the citations received in year ttt. This allows the derivation of pairwise preference labels, such as "paper iii is better than paper jjj," which form the ground truth for subsequent training stages.

In the second stage, the authors train the SCIENTIFIC JUDGE model to predict the potential impact of research ideas. A pair of papers with a citation-based preference label is fed into the judge model, which processes the input and samples a reasoning trace to reach a decision. The model outputs a predicted preference, and a correctness reward is computed against the citation-based label. This reward is defined as ri=1[y^(oi)=y]r_i = 1[\hat{y}(o_i) = y]ri=1[y^(oi)=y], where y^\hat{y}y^ is the predicted preference and yyy is the ground truth. This signal is used to update the judge model via Group Relative Policy Optimization (GRPO), enhancing its ability to evaluate scientific merit.

The final stage involves training the SCIENTIFIC THINKER model to generate high-impact research ideas. A seed paper is sampled from the dataset, and the policy π\piπ samples a group of GGG candidate ideas. The trained judge model then conducts round-robin pairwise comparisons among these candidates to evaluate their relative quality. A comparison reward, calculated as the normalized win rate, is used to update the policy. This reward is formulated as ri=1G1jis(oi,oj)r_i = \frac{1}{G-1} \sum_{j \neq i} s(o_i, o_j)ri=G11j=is(oi,oj), where s(oi,oj)s(o_i, o_j)s(oi,oj) represents the outcome of the comparison between idea iii and idea jjj. This reward signal guides the THINKER model via GRPO to produce ideas that are more likely to achieve high potential impact.

Experiment

  • Scaling experiments demonstrate that scientific judgement capability improves consistently with both increased training data and larger model sizes, with the 30B model surpassing proprietary baselines.
  • Generalization tests confirm that models trained on computer science data effectively predict paper impact across future time periods, unseen scientific fields like mathematics and biology, and different evaluation metrics such as peer-review scores.
  • Ideation training using the scientific judge as a reward model significantly enhances the quality of proposed research ideas, enabling the system to outperform base policies and compete with state-of-the-art models on future research topics.
  • Case studies illustrate that larger models successfully reason about complex factors like topic generality, institutional influence, and downstream adoption, whereas smaller models occasionally rely on misleading heuristics regarding popularity or foundationality.
  • Specialized training for scientific taste preserves general-purpose reasoning and knowledge capabilities, showing minimal degradation on standard benchmarks.

Créer de l'IA avec l'IA

De l'idée au lancement — accélérez votre développement IA avec le co-codage IA gratuit, un environnement prêt à l'emploi et le meilleur prix pour les GPU.

Codage assisté par IA
GPU prêts à l’emploi
Tarifs les plus avantageux

HyperAI Newsletters

Abonnez-vous à nos dernières mises à jour
Nous vous enverrons les dernières mises à jour de la semaine dans votre boîte de réception à neuf heures chaque lundi matin
Propulsé par MailChimp