Great scientists have strong judgement and foresight, closely tied to what we call scientific taste.
Here, we use the term to refer to the capacity to judge and propose research ideas with high potential impact.
However, most related research focuses on improving an AI scientist's executive capability, while enhancing
an AI's scientific taste remains underexplored.
We propose Reinforcement Learning from Community Feedback (RLCF), a training paradigm that uses
large-scale community signals as supervision and formulates scientific taste learning as a preference modeling
and alignment problem. For preference modeling, we train Scientific Judge on 700K field- and time-matched
pairs of high- vs. low-citation papers. For preference alignment, using Scientific Judge as a reward model,
we train Scientific Thinker to propose research ideas with high potential impact.
Experiments show that Scientific Judge outperforms strong LLM baselines such as GPT-5.2 and Gemini 3 Pro,
and generalizes to future-year test sets, unseen fields, and peer-review preference. Scientific Thinker further
proposes research ideas with higher potential impact than strong baselines. Our findings show that AI can learn
scientific taste, marking a key step toward human-level AI scientists.