Artificial intelligence might reshape the approach to science, from generating concepts and hypotheses to designing experiments, analysing data and publishing findings. However, it might also introduce risks around data integrity, reproducibility, accountability and bias. Here we provide practical guidance on how to responsibly integrate artificial intelligence into the research pipeline.
This is a preview of subscription content, access via your institution
Access options
Subscribe to this journal
Receive 12 digital issues and online access to articles
$119.00 per year
only $9.92 per issue
Buy this article
- Purchase on SpringerLink
- Instant access to the full article PDF.
USD 39.95
Prices may be subject to local taxes which are calculated during checkout
Change history
25 November 2025
A Correction to this paper has been published: https://doi.org/10.1038/s44222-025-00397-0
References
Ogunniye, G. & Hana, A. Research integrity in the era of generative AI: a perspective. UKCORI https://go.nature.com/47z7KSR (2025).
Naddaf, M. How are researchers using AI? Survey reveals pros and cons for science. Nature https://doi.org/10.1038/d41586-025-00343-5 (2025).
Resnik, D. B. & Hosseini, M. The ethics of using artificial intelligence in scientific research: new guidance needed for a new tool. AI Ethics 5, 1499–1521 (2025).
Woodhams, J., Parry, J. & Scott, J. Embracing AI with integrity. UKRIO https://go.nature.com/4oKX09m (2025).
National Academy of Medicine. An Artificial Intelligence Code of Conduct for Health and Medicine: Essential Guidance for Aligned Action. https://doi.org/10.17226/29087 (2025).
Blau et al. Protecting scientific integrity in an age of generative AI. Proc. Natl Acad. Sci. USA 121, 1–3 (2024).
Thelwall, M., Lehtisaari, M., Katsirea, I., Holmberg, K. & Zheng, E. Does ChatGPT ignore article retractions and other reliability concerns? Learn. Publ. 38, e2018 (2025).
Moons, K. G. M. et al. PROBAST+AI: an updated quality, risk of bias, and applicability assessment tool for prediction models using regression or artificial intelligence methods. BMJ 388, e082505 (2025).
Collins, G. S. et al. TRIPOD+AI statement: updated guidance for reporting clinical prediction models that use regression or machine learning methods. BMJ 385, e078378 (2024).
Acknowledgements
The authors used Zoom AI Companion during drafting sessions to assist in the summary of notes from sessions and action items for authors. The authors also used ChatGPT 5-Auto to shorten and condense draft sections and for brainstorming, followed by discussions among authors. Our workflow was organized and aligned with consideration of the research data lifecycle (for example, the Biomedical Data Lifecycle (https://datamanagement.hms.harvard.edu/)).
Author information
Authors and Affiliations
Contributions
All authors contributed to the original draft, writing of the manuscript, review, editing and supervision.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Related links
Federal sponsors: https://grants.nih.gov/grants/guide/notice-files/NOT-OD-25-132.html
Guidance from universities: https://www.harvard.edu/ai/research-resources/
Guidelines on expectations: https://osp.od.nih.gov/policies/artificial-intelligence/
Mandates on the use of AI: https://grants.nih.gov/grants/guide/notice-files/NOT-OD-25-081.html
Publishers: https://www.nature.com/nature/editorial-policies/ai
Rights and permissions
About this article
Cite this article
Rolauffs, B., Kato, R., Hart, M.L. et al. Rethinking the scientific method in the age of AI. Nat Rev Bioeng 4, 2–3 (2026). https://doi.org/10.1038/s44222-025-00386-3
Published:
Version of record:
Issue date:
DOI: https://doi.org/10.1038/s44222-025-00386-3