3 Sources
3 Sources
[1]
Artificial intelligence tools expand scientists' impact but contract science's focus - Nature
Although our analysis provides new insight into AI's impact on science, clear limitations remain. Our identification approach -- although validated by experts -- misses subtle and unmentioned forms of AI use, and our focus on natural sciences excludes important domains in which AI adoption patterns may differ. Moreover, despite consistently suggestive evidence, we cannot fully identify the causal linkage between AI adoption and scientific impact. Nevertheless, our findings demonstrate that currently attributed uses of AI in science primarily augment cognitive tasks through data processing and pattern recognition. Looking forward, these findings illuminate a critical and expansive pathway for AI development in science. To preserve collective exploration in an era of AI use, we will need to reimagine AI systems that expand not only cognitive capacity but also sensory and experimental capacity49,50, enabling and incentivizing scientists to search, select and gather new types of data from previously inaccessible domains rather than merely optimizing analysis of standing data. The history of major discoveries has been most consistently linked with new views on nature51. Expanding the scope of AI's deployment in science will be required for sustained scientific research and to stimulate new fields rather than merely automate existing ones. In this section we introduce the procedure of selecting the research papers included in our analysis. We conduct our major analyses on OpenAlex -- a scientific research database built on the foundation of the Microsoft Academic Graph (MAG). Supported by non-profit organizations, OpenAlex is continuously updated, providing a sustainable global resource for research information. As of March 2025, OpenAlex contains 265.7 million research papers, along with related data about citation, author, institution and so on. Among the massive quantity of papers in the OpenAlex dataset, we select 66,117,158 English research papers published in journals and conferences spanning from 1980 to 2025 and filter out those with incomplete titles or abstracts. We identify the scientific discipline each paper belongs to by making use of the topics noted in OpenAlex, which are extracted using a natural language processing approach that annotates titles and abstracts with Wikipedia article titles as topics sharing textual similarity. In the raw dataset, these topics form a hierarchical structure and each paper is associated with several. Adopting the 19 basic scientific disciplines in MAG, that is, art, biology, business, chemistry, computer science, economics, engineering, environmental science, geography, geology, history, materials science, mathematics, medicine, philosophy, physics, political science, psychology, and sociology, we trace along the hierarchy and determine to which disciplines each topic belongs. We note that because the original topics of one paper may be retraced to different topics, the scientific discipline of each paper may not be unique. In other words, one paper may span two or more academic disciplines, for example, chemistry and biology, which reflects the common phenomena of borderline or interdisciplinary research. Here we emphasize the adoption of AI methods in conventional natural science disciplines and exclude research developing AI methodologies themselves, separating the influence of AI on science from AI's own invention and refinement. We therefore select biology, medicine, chemistry, physics, materials science and geology as representatives of natural science disciplines, but exclude computer science and mathematics, where most works introducing and developing AI methods are published. We also exclude art, business, economics, history, philosophy, political science, psychology and sociology to focus on how AI is changing the natural sciences and career trajectories in those sciences. Our six natural science disciplines include the majority of OpenAlex articles, resulting in 41,298,433 papers, containing 18,392,040 in biology, 4,209,771 in chemistry and 2,380,666 in geology, 4,755,717 in materials science, 24,315,342 in medicine and 5,138,488 in physics. The selected disciplines cover various dimensions of natural science, representing a broad view of scientific research as a whole. We divide the history of AI development into three key eras: the traditional machine learning era (1980-2014), the deep learning era (2015-2022) and the generative AI era (2023 to present). We consider 1980 as the start of the traditional machine learning era because several landmark works were published in the 1980s, such as the back-propagating method. We regard the deep learning era to have begun in 2015, as indicated by breakthroughs such as ResNet, which enabled the training of ultra-deep neural networks, revolutionizing fields such as computer vision and speech recognition. Finally, we define the generative AI era as beginning in 2023, following the publication of ChatGPT -- the first widely used large language model -- in December 2022. This marked the advent of large-scale transformer-based models capable of strong generalized performance across a wide range of tasks, sparking new applications in natural language processing and beyond. Each of these transitions was driven by advances in algorithms, computational power and data availability, substantially expanding the capabilities and scope of AI for science. Insofar as both a paper's title and abstract contain important information about its content, we independently train two separate models on the basis of paper titles and abstracts, and then integrate the two models into an ensembled one by averaging their outputs. The structure of our natural language processing model for paper identification consists of two parts. The backbone network is a twelve-layer BERT model with twelve attention heads in each layer, and the sequence classification head is a linear layer with a two-dimensional output atop the BERT output. We normalize the two-dimensional output with a softmax function and obtain the probability that the paper involves AI-assistance. We use the BERT model called BERT-base-uncased from Hugging Face, which is pre-trained with a large-scale general corpus, and set the maximum length of tokenization to be 16 for titles and 256 for abstracts. We design a two-stage fine-tuning process with training and validation sets, which we extracted from the OpenAlex dataset, to transfer the pre-trained model to our paper identification task. The construction of positive and negative data is different between the two stages. In both stages, we randomly split the positive and negative data into 90% and 10% sets, which correspond to training and validation sets, respectively. We use the training set for model training and use the validation set to select the optimal model. As the numbers of positive and negative cases are unbalanced, we use the bootstrap sample technique on positive cases to balance its number with negative cases at both stages. In the first stage, we construct relatively coarse positive data, only considering eight typical AI journals and conferences, including Nature Machine Intelligence, Machine Learning, Artificial Intelligence, Journal of Machine Learning Research (JMLR), International Conference on Machine Learning (ICML), International Conference on Learning Representations (ICLR) and the AAAI Conference on Artificial Intelligence and International Joint Conference on Artificial Intelligence (IJCAI). Among the papers belonging to our chosen six disciplines, we extract all papers published in these venues as positive cases and randomly sample 1% of the remaining papers in our six chosen natural science fields as negative cases, resulting in 26,165 positive and 291,035 negative cases. We fine-tune the pre-trained model for 30 epochs on the training set and select the optimal model according to the F1-score on the validation set. In the second stage, we construct more precise positive data on the basis of the optimal model obtained in the first stage. We identify papers in the whole OpenAlex dataset and aggregate the results for each venue, obtaining the probability that each venue in OpenAlex is an AI venue by averaging the AI probability for all papers within it. We then select the venues with >80% AI probability and >100 papers as AI venues. We also incorporate venues with 'machine learning' or 'artificial intelligence' in their names. In papers belonging to our six chosen disciplines, we extract all papers published in the selected AI venues as positive cases and randomly sample 1% of those remaining as negative cases, resulting in 31,311 positive and 231,258 negative cases. We then fine-tune the obtained optimal model in the first stage for another 30 epochs with the new training set and select the best model according to F1-score on the new validation set. Finally, we use optimal ensemble models during both stages to identify all papers that use AI to support natural science research from the selected representative natural science disciplines. We arbitrarily sample 220 papers (110 papers × 2 groups) from each of the six disciplines, resulting in twelve paper groups in total. We enlisted twelve experts with abundant AI research experience (Supplementary Table 1) and assigned three different groups of papers to each. Without revealing the classification results obtained from the BERT model, we queried our experts on whether each paper was an AI paper. In this way, each paper is repeatedly labelled by three distinct experts, and we evaluate the consistency among these experts on the basis of Fleiss's κ (refs. ), which is an unsupervised measurement for assessing the agreement between independent raters. Having confirmed consensus among our experts, we draw the final expert label of each paper from the three experts according to the principle of the minority obeying the majority. We regard the expert labels as ground truth and validate the result of our BERT model against them with the F1-score, which is a supervised measurement of accuracy. Here we define the project leader as the last author of a research paper, in alignment with conventions established by previous studies. To ensure that in most papers, the last author represents the project leader, we examine the fraction of papers that list authors following alphabetical order. First, we directly traverse all selected papers and obtain the prevalence of papers listing authors in alphabetical order, which ranges from 14.87% in materials science to 22.15% in geology. Nevertheless, it is difficult to distinguish whether these papers actually intended to list the authors in alphabetical order or according to their roles, which unintentionally fall in alphabetical order. The latter situation is more likely to occur when there are fewer authors (two or three). To tackle this analytical challenge, we determine the fraction of unintended alphabetical author lists through a Monte Carlo method. We generate ten randomly shuffled copies of the author list for each paper and find that from 13.82% (materials science, σ = 0.02) to 20.28% (geology, σ = 0.03) of papers have alphabetically listed authors among the random author lists. This indicates the proportion of 'unintended' alphabetical author lists, and we can derive the actual fraction of papers with intentionally alphabetical author lists by the difference between the above two sets of statistical results. The actual fraction obtained illustrates that only 1.58% of papers across all disciplines intentionally list the authors in alphabetical order (Supplementary Table 12) and therefore, we can, with negligible interference, assume that we can identify last authors as team leaders. The OpenAlex dataset incorporates a well-designed author name disambiguation mechanism, which uses an XGBoost model to predict the likelihood that two authors are the same on the basis of features such as their institutions, co-authors and citations, and then applies a custom, ORCID-anchored clustering process to group their works, assigning a unique ID to each author. Simply using unique IDs, we are able to track a large number of authors at the same time, where we depict an individual scientist's career trajectory using a role transition model (Extended Data Fig. 4a) and extract the role transition trajectories for scientists. First we traverse all selected papers in the six disciplines and extract all the scientists involved in any of these papers. Then, for each individual scientist, we extract all papers in which they have been involved and record the time of their first publication in any role, the time of their first publication as team leader (if ever), and the time of their last publication. We then filter out scientists whose publication records span only a single year. We also filter out those who directly start as established scientists leading research teams without a role transition from junior scientists. Finally, we detect the time that each scientist abandons academic publishing. Considering that one scientist may not publish papers continuously every year, we cannot regard them as having left academia on the basis of their absence in the published record for a single year. We therefore follow the settings used in previous work to use a threshold of three years and regard scientists who have no more publications after 2022 as having exited academia, whereas those who still publish papers after 2022 are considered to have an unclear ultimate status and are excluded from the analysis. Finally, we obtain 2,282,029 scientists in the six disciplines with complete role transition trajectories. We also classify them into AI and non-AI scientists according to whether they have published AI-augmented papers. Moreover, by analysing author contribution statements collected in previous studies, we further validate our detection results by examining changes in scientists' self-reported contributions throughout their careers (Extended Data Fig. 4b). Results indicate that junior scientists primarily engage in technical tasks, such as conducting experiments and analysing data, and less in conceptual tasks, such as conceiving ideas and writing papers. Nevertheless, the proportion of conceptual work significantly rises (P < 0.01 and df = 1 in a Cochran-Armitage test) during their tenure as junior scientists, reaching saturation at a high level (60% or more) on transition to becoming established scientists. This finding validates our definition of role transition by demonstrating a shift in the nature of scientists' contributions from participating in research projects to leading them. To obtain a more precise quantification of how much AI accelerates the career development of junior scientists, we use a general birth-death model. This type of stochastic process model depicts the dynamic evolution of a population as members join and exit. In our context, it models the role transitions of junior scientists. Specifically, we use two separate birth-death models for junior scientists who eventually become established and for those who leave academia, respectively. Here, 'birth' processes refer to the entry of junior scientists into academic publishing, and 'death' processes symbolize their transition out of the junior stage, either by becoming established scientists or quitting academia. As the entry and exit of each junior scientist are independent from one another, we use Poisson processes to model 'birth' (entry) and 'death' (exit) events, respectively. The Poisson process is a typical stochastic process model for describing the occurrence of random events that are independent of each other. The mathematical formula of the Poisson process is: where N(t) denotes the number of random events that happened before time t, and λ is the parameter of the Poisson process, depicting the happening rate of random events. We consider a birth-death model in which birth and death dynamics are both Poisson processes, and rate parameters are μ and ω, respectively. Through mathematical derivation, we conclude that the duration time t from birth to death follows an exponential distribution with the parameter ω - μ, where the exact form of the probability density function is: We consider the difference between the two rate parameters ω - μ as a whole and fit it with a single parameter λ. The transition time for junior scientists to become established scientists or leave academia then follows the exponential distribution: and the corresponding survival function is Hence the average transition time is the conditional expectation of the distribution defined as follows: We fit the role transition time of scientists with the aforementioned exponential distribution, thereby determining the respective values of λ for AI-adopting junior scientists and their non-AI counterparts. Guided by the underlying mechanism of junior scientists' career development incorporated within the birth-death model, expectations from the model offer a more accurate estimate of the average role transition time. To assess the knowledge extent of a set of research papers within their high-dimensional embeddings we first compute the centroid as the mean of their vector locations: Next, we compute the Euclidean distance from each embedding to the centroid, where the knowledge extent (KE) of the set of papers is defined as the maximum distance or 'diameter' of the vector space covered: We note that Euclidean distance is highly correlated with the cosine and related angular distances. In practice, the number of AI and non-AI papers in each domain differs considerably, introducing bias to the measurement of knowledge extent. To address this issue, we build on past work about cognitive extent, which is a measure of the breadth of a scientific field's cognitive territory, and is quantified by the number of unique phrases -- as a proxy for scientific concepts -- found within a sampled batch of papers of a given size. For each domain, we randomly sample 1,000 papers from both AI and non-AI categories, compute their respective knowledge extent, and repeat this process 1,000 times. By comparing knowledge extent values across these 1,000 random samples, we ensure that the number of AI and non-AI papers is balanced, making our knowledge extent results comparable. To measure how much knowledge space can be derived from each original paper, we calculate the knowledge extent of 'paper families', that is, a focal paper and its follow-on citations. Focusing on an original research paper ϕ, which corresponds to a high-dimensional embedding vector , we extract all n research papers that cite this original paper. These papers are sorted chronologically by publication date, from earliest to most recent. The corresponding high-dimensional embeddings of these sorted papers are: Thereby, we calculate knowledge extent covered by the 'paper family' consisting of the original paper ϕ and the first n follow-on papers, citing it (1 ≤ n ≤ n) as: To quantify how frequently citations of the same original paper interact with each other, we design a metric called follow-on engagement, building on previous work. For an original paper with n citations, there are at most possible citations among these n citing papers if everyone cites all papers published earlier than their own. We then count how many times these n citing papers actually cite one another, denoted as k. Our metric for follow-on engagement (EG) is calculated as the ratio of actual to maximum possible citations: This metric helps quantify the degree of interactions and collaboration among papers that cite the same original work. Past work has demonstrated a positive association between the ambiguity of a focal work and follow-on engagement. Further information on research design is available in the Nature Portfolio Reporting Summary linked to this article.
[2]
AI has supercharged scientists -- but may have shrunk science
Analysis of 41 million papers finds that although AI expands individual impact, it narrows collective scientific exploration As artificial intelligence tools such as ChatGPT gain footholds across companies and universities, a familiar refrain is hard to escape: AI won't replace you, but someone using AI might. A paper published today in Nature suggests this divide is already creating winners and laggards in the natural sciences. In the largest analysis of its kind so far, researchers find that scientists embracing any type of AI -- going all the way back to early machine learning methods -- consistently make the biggest professional strides. AI adopters have published three times more papers, received five times more citations, and reach leadership roles faster than their AI-free peers. But science as a whole is paying the price, the study suggests. Not only is AI-driven work prone to circling the same crowded problems, but it also leads to a less interconnected scientific literature, with fewer studies engaging with and building on one another. "I was really amazed by the scale and scope of this analysis," says Yian Yin, a computational social scientist at Cornell University who has studied the impact of large language models (LLMs) on scientific research. "The diversity of AI tools and very different ways we use AI in scientific research makes it extremely hard to quantify these patterns." These results should set off "loud alarm bells" for the community at large, adds Lisa Messeri, a sociocultural anthropologist at Yale University. "Science is nothing but a collective endeavor," she says. "There needs to be some deep reckoning with what we do with a tool that benefits individuals but destroys science." To uncover these trends, researchers began with more than 41 million papers published from 1980 to 2025 across biology, medicine, chemistry, physics, materials science, and geology. First, they faced a major hurdle: figuring out which papers used AI, a category that spans everything from early machine learning to today's LLMs. "This is something that people have been trying to figure out for years, if not for decades," Yin says. The team's solution was, fittingly, to use AI itself. The researchers trained a language model to scan titles and abstracts and flag papers that likely relied on AI tools, identifying about 310,000 such papers in the data set. Human experts then reviewed samples of the results and confirmed the model was about as accurate as a human reviewer. With that subset of papers, the researchers could then measure AI's impact on the scientific ecosystem. Across the three major eras of AI -- machine learning from 1980 to 2014, deep learning from 2016 to 2022, and generative AI from 2023 onward -- papers that used AI drew nearly twice as many citations per year as those that did not. Scientists who adopted AI also published 3.02 times as many papers and received 4.84 times as many citations over their careers. Benefits extended to career trajectories, too. Zooming in on 2 million of the researchers in the data set, the team found that junior scientists who used AI were less likely to drop out of academia and more likely to become established research leaders, doing so nearly 1.5 years earlier than their peers who hadn't. But what was good for individuals wasn't good for science. When the researchers looked at the overall spread of topics covered by AI-driven research, they found that AI papers covered 4.6% less territory than conventional scientific studies. This clustering, the team hypothesizes, results from a feedback loop: Popular problems motivate the creation of massive data sets, those data sets make the use of AI tools appealing, and advances made using AI tools attract more scientists to the same problems. "We're like pack animals," says study co-author James Evans, a computational social scientist at the University of Chicago. That crowding also shows up in the links between papers. In many fields, new ideas grow through dense networks of papers that cite one another, refine methods, and launch new lines of research. But AI-driven papers spawned 22% less engagement across all the natural sciences disciplines. Instead, they tended to orbit a small number of superstar papers, with fewer than one-quarter of papers receiving 80% of the citations. "When your attention is attracted by star papers like [the protein folding model] AlphaFold, all you're thinking is how you can build on AlphaFold and beat other people to doing it," says Tsinghua University co-author Fengli Xu. "But if we all climb the same mountains, then there are a lot of fields we are not exploring." "Science is seeing a degree of disruption that is rare," says Dashun Wang, who researches the science of science at Northwestern University. The rapid rise of generative AI -- which is reshaping research workflows faster than many scientific institutions can keep up -- only makes the stakes higher and the future shape of science less certain, he says. But the narrowing of science may still be reversible. One way to push back, says Zhicheng Lin, a psychologist at Yonsei University who studies the science of science, is to build better and larger data sets in fields that haven't yet made much use of AI. "We are not going to improve science by forcing a shift away from data-heavy approaches," he says. "A brighter future involves making data more abundant across more domains." Further down the line, future AI systems should also evolve beyond crunching data into autonomous agents capable of scientific creativity, which could expand science's horizons again, says study co-author Yong Li, who studies AI and the science of science at Tsinghua. Until then, Evans says, the scientific community must reckon with how these tools have affected incentives across the board. "I don't think this is how AI has to shape science," he says. "We want a world in which AI-enhanced work, which is getting increased funding and increasing in rate, is generating new fields -- rather than just turning the thumbscrews on old questions."
[3]
AI tools boost individual scientists but could limit research as a whole
Artificial intelligence is influencing many aspects of society, including science. Writing in Nature, Hao et al. report a paradox: the adoption of AI tools in the natural sciences expands scientists' impact but narrows the set of domains that research is carried out in. The authors examined more than 41 million papers, roughly 311,000 of which had been augmented by AI in some way -- through the use of machine-learning methods or generative AI, for example. They find that scientists who conduct AI-augmented research publish more papers, are cited more often and progress faster in their careers than those who do not, but that AI automates established fields rather than supporting the exploration of new ones. This raises questions and concerns regarding the potential impact of AI tools on scientists and on science as a whole.
Share
Share
Copy Link
A groundbreaking analysis of over 41 million research papers reveals a stark paradox in modern science. While artificial intelligence tools dramatically boost individual researchers' productivity and career success, they simultaneously narrow the collective scope of scientific exploration. Scientists using AI publish three times more papers and receive five times more citations, yet AI-driven research covers 4.6% less scientific territory than traditional studies.
A comprehensive study published in Nature
1
analyzing more than 41 million research papers from 1980 to 2025 reveals that artificial intelligence (AI) has fundamentally altered the landscape of scientific research. The analysis, which examined approximately 311,000 AI-augmented papers across biology, medicine, chemistry, physics, materials science, and geology, demonstrates that scientists adopting AI tools achieve remarkable individual success2
. Researchers leveraging machine learning, deep learning, or generative AI publish 3.02 times more research papers and receive 4.84 times more citations over their careers compared to peers who avoid these tools.
Source: Nature
The impact of AI tools on scientific research extends beyond publication metrics to reshape career trajectories entirely. Junior scientists who embrace AI-driven research are significantly less likely to drop out of academia and reach leadership positions nearly 1.5 years earlier than their counterparts
2
. Papers utilizing AI attract nearly twice as many citations per year across three major eras of AI development: traditional machine learning from 1980 to 2014, deep learning from 2015 to 2022, and generative AI from 2023 onward. These benefits for individual researchers create a clear competitive advantage in an increasingly AI-saturated academic environment.Despite these individual gains, the study reveals a troubling consequence for science as a collective endeavor. AI-augmented science covers 4.6% less scientific territory than conventional studies, indicating that the impact of AI is narrowing scientific exploration rather than expanding it
2
. This contraction stems from a self-reinforcing cycle where popular problems generate massive datasets, those datasets make AI tools appealing, and advances using AI attract more scientists to the same crowded domains. "We're like pack animals," explains study co-author James Evans from the University of Chicago, describing how researchers cluster around the same high-profile challenges.The scientific literature itself reflects this narrowing focus. AI-driven research spawned 22% less engagement across natural sciences disciplines, with papers tending to orbit a small number of superstar publications rather than forming dense, interconnected networks of ideas
2
. Fewer than one-quarter of papers receive 80% of citations, creating a winner-take-all dynamic that discourages exploration of less-traveled research paths. This pattern suggests that AI augmenting science primarily through data processing and pattern recognition may be automating established fields rather than stimulating new ones.Related Stories
The findings illuminate critical questions about the long-term trajectory of scientific discovery. Researchers used OpenAlex, a database containing 265.7 million research papers as of March 2025, to identify AI adoption patterns
1
. To detect which papers employed AI, the team trained a language model to scan titles and abstracts, achieving accuracy comparable to human reviewers. This methodology itself reflects how deeply AI has penetrated research workflows, even in studies examining AI's influence.Experts warn that these trends demand urgent attention. "Science is nothing but a collective endeavor," notes Lisa Messeri, a sociocultural anthropologist at Yale University, adding that the field needs "some deep reckoning with what we do with a tool that benefits individuals but destroys science"
2
. The study suggests that preserving collective exploration will require reimagining AI systems that expand not only cognitive capacity but also sensory and experimental capacity, enabling scientists to gather new types of data from previously inaccessible domains rather than merely optimizing analysis of existing datasets1
.The rapid rise of generative AI has accelerated these dynamics faster than scientific institutions can adapt. As Dashun Wang from Northwestern University observes, "Science is seeing a degree of disruption that is rare"
2
. The history of major discoveries has consistently linked breakthroughs with new perspectives on nature, suggesting that expanding the scope of AI's deployment beyond pattern recognition will be essential for sustained scientific progress. Researchers and institutions must now balance the undeniable advantages AI provides to individual career advancement against the risk of creating an increasingly homogenized research landscape that favors optimization over exploration3
.Summarized by
Navi
[1]
12 Oct 2024•Science and Research

05 Aug 2025•Science and Research

06 Jan 2025•Science and Research

1
Policy and Regulation

2
Technology

3
Policy and Regulation
