Skip to main content

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

Advertisement

Scientific Reports
  • View all journals
  • Search
  • My Account Login
  • Content Explore content
  • About the journal
  • Publish with us
  • Sign up for alerts
  • RSS feed
  1. nature
  2. scientific reports
  3. articles
  4. article
Benchmarking action recognition models for self-harm detection in studio and real-world datasets
Download PDF
Download PDF
  • Article
  • Open access
  • Published: 31 January 2026

Benchmarking action recognition models for self-harm detection in studio and real-world datasets

  • Kanghee Lee1,2 na1,
  • Doohee Lee1,2 na1,
  • Hyun-Sik Ham1,
  • Hee-Cheol Kim1,
  • Yourack Lee3,
  • Hyun-Ghang Jeong4 &
  • …
  • Hyun-Soo Choi1,5 

Scientific Reports , Article number:  (2026) Cite this article

We are providing an unedited version of this manuscript to give early access to its findings. Before final publication, the manuscript will undergo further editing. Please note there may be errors present which affect the content, and all legal disclaimers apply.

Subjects

  • Engineering
  • Mathematics and computing

Abstract

The development of effective automated systems to prevent patient self-harm in psychiatric wards is severely hampered by a scarcity of realistic training data. To address this critical gap, this study introduces a new public dataset of 1120 videos simulating cutting actions in a controlled studio environment, alongside a validation set of 118 real-world videos from secure wards that include more diverse behaviors such as picking and scratching. We conducted a comprehensive benchmark of state-of-the-art action recognition models, including both convolution-based and transformer-based architectures, to evaluate their performance and generalizability from simulated to real-world conditions. Our results reveal a significant “sim-to-real” gap: while the top-performing model, VideoMAEv2, achieved a high F1 score of 0.65 on the simulated data using 7-fold LOAO cross-validation, its performance degraded to a mean F1 score of 0.61 on the real-world data. This performance drop is attributed to the models’ inability to generalize from the uniform, simulated actions to the diverse and often occluded behaviors observed in authentic clinical settings. By providing a foundational dataset, a systematic benchmark, and a qualitative analysis of model failure points, this study quantitatively demonstrates the limitations of current approaches. Our findings underscore the urgent need for more diverse data and advanced approaches to develop robust technologies that can enhance patient safety.

Similar content being viewed by others

A novel YOLO LSTM approach for enhanced human action recognition in video sequences

Article Open access 16 May 2025

Attention-based bidirectional-long short-term memory for abnormal human activity detection

Article Open access 02 September 2023

Generalizability of clinical prediction models in mental health

Article Open access 19 March 2025

Data availability

The studio-based self-harm dataset generated and analyzed during the current study is publicly available in the ZV_Self-harm-Dataset repository, accessible at https://github.com/zv-ai/ZV_Self-harm-Dataset.

Code availability

Code is available upon request by the corresponding author.

References

  1. Allison, S., Bastiampillai, T., Looi, J. C., Kisely, S. R. & Lakra, V. The new world mental health report: Believing impossible things. Aust. Psychiatry 31, 182–185. https://doi.org/10.1177/10398562231154806 (2023).

  2. Kool, N. & Jaspers, A. Self-harm on a closed psychiatric ward. Eur. Psychiatry 65, S663–S663. https://doi.org/10.1192/j.eurpsy.2022.1703 (2022).

    Google Scholar 

  3. Kuehne, H., Jhuang, H., Garrote, E., Poggio, T. & Serre, T. HMDB: A large video database for human motion recognition. In Proceedings of the International Conference on Computer Vision (ICCV) (2011).

  4. Soomro, K., Zamir, A. R. & Shah, M. Ucf101: A dataset of 101 human actions classes from videos in the wild. arXiv:1212.0402 (2012).

  5. Kay, W. et al. The kinetics human action video dataset. arXiv:1705.06950 (2017).

  6. Donahue, J. et al. Long-term recurrent convolutional networks for visual recognition and description. arXiv:1411.4389 (2016).

  7. Tran, D., Bourdev, L., Fergus, R., Torresani, L. & Paluri, M. Learning spatiotemporal features with 3D convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision. 4489–4497 (2015).

  8. Bertasius, G., Wang, H. & Torresani, L. Is space-time attention all you need for video understanding? arXiv:2102.05095 (2021).

  9. Li, K. et al. Uniformerv2: Spatiotemporal learning by arming image VITS with video uniformer. arXiv:2211.09552 (2022).

  10. Dosovitskiy, A. et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv:2010.11929. (2021).

  11. Liang, Z. et al. Nssi-net: Multi-concept generative adversarial network for non-suicidal self-injury detection using high-dimensional EEG signals in a semi-supervised learning framework. arXiv:2410.12159 (2024).

  12. Krizhevsky, A., Sutskever, I. & Hinton, G. E. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems (Pereira, F., Burges, C., Bottou, L. & Weinberger, K. eds.) . Vol. 25 (Curran Associates, Inc., 2012).

  13. Scherr, S., Arendt, F., Frissen, T. & M, J. O. Detecting intentional self-harm on Instagram: Development, testing, and validation of an automatic image-recognition algorithm to discover cutting-related posts. Soc. Sci. Comput. Rev. 38, 673–685. https://doi.org/10.1177/0894439319836389 (2020).

  14. Yang, G. et al. Detection of non-suicidal self-injury based on spatiotemporal features of indoor activities. IET Biometrics 12, 91–101. https://doi.org/10.1049/bme2.12110 (2023).

  15. Gardner, K. J. et al. The significance of site of cut in self-harm in young people. J. Affect. Disord. 266, 603–609 (2020).

    Google Scholar 

  16. Corporation, C. Computer Vision Annotation Tool (CVAT). https://doi.org/10.5281/zenodo.8416684 (2023).

  17. Wang, L. et al. Temporal segment networks for action recognition in videos (2017). arXiv:1705.02953.

  18. Feichtenhofer, C., Fan, H., Malik, J. & He, K. Slowfast networks for video recognition. arXiv:1812.03982 (2019).

  19. Wang, L. et al. Videomae v2: Scaling video masked autoencoders with dual masking. arXiv:2303.16727 (2023).

  20. Al-lahham, A., Tastan, N., Zaheer, Z. & Nandakumar, K. A coarse-to-fine pseudo-labeling (c2fpl) framework for unsupervised video anomaly detection. arXiv:2310.17650 (2023).

  21. Zhang, H., Li, X. & Bing, L. Video-llama: An instruction-tuned audio-visual language model for video understanding. arXiv:2306.02858 (2023).

  22. OpenAIet al. Gpt-4o system card. arXiv:2410.21276 (2024).

Download references

Author information

Author notes
  1. Kanghee Lee and Doohee Lee contributed equally to this work.

Authors and Affiliations

  1. Department of Research and Development, Ziovision Co., Ltd., Chuncheon, 24341, Gangwon, Republic of Korea

    Kanghee Lee, Doohee Lee, Hyun-Sik Ham, Hee-Cheol Kim & Hyun-Soo Choi

  2. Department of Computer and Communications Engineering, College of IT, Kangwon National University, Chuncheon, 24341, Gangwon, Republic of Korea

    Kanghee Lee & Doohee Lee

  3. Biomedical Research Center, Korea University Guro Hospital, Seoul, 08308, Republic of Korea

    Yourack Lee

  4. Department of Psychiatry, Korea University Guro Hospital, Korea University College of Medicine, Seoul, 08308, Republic of Korea

    Hyun-Ghang Jeong

  5. Department of Computer Science and Engineering, Seoul National University of Science and Technology, Seoul, 01811, Republic of Korea

    Hyun-Soo Choi

Authors
  1. Kanghee Lee
    View author publications

    Search author on:PubMed Google Scholar

  2. Doohee Lee
    View author publications

    Search author on:PubMed Google Scholar

  3. Hyun-Sik Ham
    View author publications

    Search author on:PubMed Google Scholar

  4. Hee-Cheol Kim
    View author publications

    Search author on:PubMed Google Scholar

  5. Yourack Lee
    View author publications

    Search author on:PubMed Google Scholar

  6. Hyun-Ghang Jeong
    View author publications

    Search author on:PubMed Google Scholar

  7. Hyun-Soo Choi
    View author publications

    Search author on:PubMed Google Scholar

Contributions

All authors made significant contributions to this study and approved the final manuscript. KL and DL contributed to the research by conducting experiments and drafting the original manuscript. HSH and HCK analyzed the data and provided a revision of the manuscript YL contributed to the collection and extraction of self-harm data and to the revision of the manuscript. HGJ contributed to the understanding of self-harm behaviors and psychiatric wards, provided supervision for the research, and contributed to the revision of the manuscript. HSC contributed to the design and supervision of the research and to the critical revision of the manuscript.

Corresponding authors

Correspondence to Hyun-Ghang Jeong or Hyun-Soo Choi.

Ethics declarations

Competing interests

The authors declare no competing interests.

Ethics approval and consent to participate

This study was conducted in adherence to the Declaration of Helsinki and was approved by the International Review Board of Korea University Hospital (2022GR0511). While the IRB waived the general requirement for participant consent, written informed consent for participation and data use was specifically obtained from all actors involved in the studio dataset prior to any recording.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Lee, K., Lee, D., Ham, HS. et al. Benchmarking action recognition models for self-harm detection in studio and real-world datasets. Sci Rep (2026). https://doi.org/10.1038/s41598-026-36999-w

Download citation

  • Received: 22 September 2025

  • Accepted: 19 January 2026

  • Published: 31 January 2026

  • DOI: https://doi.org/10.1038/s41598-026-36999-w

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

Download PDF

Advertisement

Explore content

  • Research articles
  • News & Comment
  • Collections
  • Subjects
  • Follow us on Facebook
  • Follow us on Twitter
  • Sign up for alerts
  • RSS feed

About the journal

  • About Scientific Reports
  • Contact
  • Journal policies
  • Guide to referees
  • Calls for Papers
  • Editor's Choice
  • Journal highlights
  • Open Access Fees and Funding

Publish with us

  • For authors
  • Language editing services
  • Open access funding
  • Submit manuscript

Search

Advanced search

Quick links

  • Explore articles by subject
  • Find a job
  • Guide to authors
  • Editorial policies

Scientific Reports (Sci Rep)

ISSN 2045-2322 (online)

nature.com sitemap

About Nature Portfolio

  • About us
  • Press releases
  • Press office
  • Contact us

Discover content

  • Journals A-Z
  • Articles by subject
  • protocols.io
  • Nature Index

Publishing policies

  • Nature portfolio policies
  • Open access

Author & Researcher services

  • Reprints & permissions
  • Research data
  • Language editing
  • Scientific editing
  • Nature Masterclasses
  • Research Solutions

Libraries & institutions

  • Librarian service & tools
  • Librarian portal
  • Open research
  • Recommend to library

Advertising & partnerships

  • Advertising
  • Partnerships & Services
  • Media kits
  • Branded content

Professional development

  • Nature Awards
  • Nature Careers
  • Nature Conferences

Regional websites

  • Nature Africa
  • Nature China
  • Nature India
  • Nature Japan
  • Nature Middle East
  • Privacy Policy
  • Use of cookies
  • Legal notice
  • Accessibility statement
  • Terms & Conditions
  • Your US state privacy rights
Springer Nature

© 2026 Springer Nature Limited

Nature Briefing AI and Robotics

Sign up for the Nature Briefing: AI and Robotics newsletter — what matters in AI and robotics research, free to your inbox weekly.

Get the most important science stories of the day, free in your inbox. Sign up for Nature Briefing: AI and Robotics